Overview
This page provides an overview of the DeGirum Orca AI accelerator, describing its performance characteristics, support for pruned models, dedicated DRAM feature, and its flexible architecture.
Last updated
Was this helpful?
This page provides an overview of the DeGirum Orca AI accelerator, describing its performance characteristics, support for pruned models, dedicated DRAM feature, and its flexible architecture.
Last updated
Was this helpful?
DeGirumĀ® Orca is a flexible, efficient, and cost-effective AI accelerator. It helps developers build feature-rich edge solutions while staying within power and cost constraints.
Orca's efficient architecture delivers strong real-world performance. A single Orca can handle multiple input streams and several ML models. See our for performance details.
Processing pruned models effectively boosts compute and bandwidth resources, letting you run larger, more accurate models in real time at the edge.
Dedicated DRAM helps applications quickly switch between ML models without lengthy transfers from the host. This reduces model-switching delays and is especially helpful when your application needs to change models often, such as in image or speech recognition scenarios.
Orca's flexible architecture supports both int8 and float32 precision, so you can choose the format that best fits your use case and optimize performance, accuracy, and power consumption.