Inference with cloud models
Run inference on a local AI server while fetching models from DeGirum’s public cloud zoo—ideal for hybrid setups where compute is local, but model access is remote.
Estimated read time: 2 minutes
This setup runs inference on local hardware (e.g., inference_host_address="localhost" or inference_host_address="<host_ip>:<port>") while fetching models from the cloud zoo (zoo_url="degirum/hailo"). It's useful when compute is local, but you want to pull models directly from DeGirum's public AI Hub.
Client and AI Server on the same host
In this case, the AI Server and client application run on the same machine.
To start the server, run:
degirum serverYou should see output similar to:
DeGirum asio server is started at TCP port 8778
Local model zoo is served from '.' directory.
Press Enter to stop the serverThe server runs until you press ENTER in the terminal. By default, it listens on TCP port 8778. To specify a different port, use the --port argument.
degirum server --port <your_port>Example ModelSpec
This ModelSpec runs inference on a local AI Server using a model fetched from DeGirum's public Hailo model zoo:
# Example ModelSpec
model_spec = ModelSpec(
model_name="yolov8n_coco--640x640_quant_hailort_multidevice_1",
,
,
model_properties={"device_type": ["HAILORT/HAILO8L", "HAILORT/HAILO8"]}
)zoo_url="degirum/hailo": loads the model from DeGirum's public Hailo model zoo.inference_host_address="localhost:8778": runs inference using Hailo devices managed by a local AI server.
Client and AI Server on different hosts
You can also run the AI server on a remote host. On that remote machine, launch the server:
degirum serverExpected output:
DeGirum asio server is started at TCP port 8778
Local model zoo is served from '.' directory.
Press Enter to stop the serverThe server runs until you press ENTER in the terminal. By default, it listens on TCP port 8778. To specify a different port, use the --port argument:
degirum server --port <your_port>Example ModelSpec
This ModelSpec runs inference on a remote AI Server using a model fetched from DeGirum's public Hailo model zoo.
# Example ModelSpec
model_spec = ModelSpec(
model_name="yolov8n_coco--640x640_quant_hailort_multidevice_1",
,
,
model_properties={"device_type": ["HAILORT/HAILO8L", "HAILORT/HAILO8"]}
)zoo_url="degirum/hailo": fetches the model from DeGirum's public Hailo model zoo.inference_host_address="<host_ip>:<port>": runs inference using Hailo devices managed by the remote AI server.
Last updated
Was this helpful?

