Inference setup
PySDK gives you flexibility in where models are stored and where inferences run. This page walks through common setups (cloud, local, and hybrid) so you can choose what fits your workflow.
Estimated read time: 2 minutes
Inference setup—pick your scenario
This page focuses on two ModelSpec parameters:
zoo_url: where the model is stored (cloud vs. local)inference_host_address: where the model runs ("@cloud"vs."@local")
Pick the setup that matches where you want inference to run and where your models live.
Cloud inference
Cloud zoo → cloud runtime
Use when: You want zero local setup—everything runs in the cloud.
from degirum_tools import ModelSpec
spec = ModelSpec(
model_name="yolov8n_coco--640x640_quant_hailort_multidevice_1",
zoo_url="degirum/hailo", # cloud model zoo
inference_host_address="@local", # inference executes on your machine
model_properties={"device_type": ["HAILORT/HAILO8L", "HAILORT/HAILO8"]},
)
model = spec.load_model()Local inference with cloud zoo
Cloud zoo → local runtime
Use when: You have local hardware but prefer to fetch models from the cloud.
Local inference with local zoo
Local zoo → local runtime
Use when: You want offline operation and predictable model behavior. Use ModelSpec.ensure_local() to download and store the model locally, then switch the ModelSpec to local before loading.
Last updated
Was this helpful?

