Skip to main content
Tooledge-aiv1.17

ONNX Runtime Mobile

by Microsoft · free · Last verified 2026-03-17

ONNX Runtime Mobile is Microsoft's high-performance inference engine optimized for mobile and edge devices, enabling deployment of models from any ONNX-compatible training framework. It provides hardware-accelerated inference via NNAPI, Core ML, and XNNPACK execution providers.

https://onnxruntime.ai/docs/tutorials/mobile/
B+
B+Good
Adoption: AQuality: AFreshness: ACitations: AEngagement: F

Specifications

License
MIT
Pricing
free
Capabilities
cross-framework inference, hardware acceleration, model optimization, operator pruning, quantization
Integrations
Android NNAPI, Apple Core ML, XNNPACK, DirectML, TensorRT
Use Cases
mobile model deployment, edge inference pipelines, cross-platform ML serving, real-time computer vision
API Available
No
SDK Languages
python, c++, c#, java, javascript
Deployment
edge, mobile, embedded, cloud
Rate Limits
No limits — fully on-device
Data Privacy
On-device processing; no cloud dependency required
Tags
edge, onnx, mobile, inference, cross-platform, iot
Added
2026-03-17
Completeness
100%

Index Score

70.4
Adoption
82
Quality
88
Freshness
82
Citations
80
Engagement
0

Explore the full AI ecosystem on Agents as a Service