ONNX Runtime Mobile
by Microsoft · free · Last verified 2026-03-17
ONNX Runtime Mobile is Microsoft's high-performance inference engine optimized for mobile and edge devices, enabling deployment of models from any ONNX-compatible training framework. It provides hardware-accelerated inference via NNAPI, Core ML, and XNNPACK execution providers.
https://onnxruntime.ai/docs/tutorials/mobile/ ↗B+
B+—Good
Adoption: AQuality: AFreshness: ACitations: AEngagement: F
Specifications
- License
- MIT
- Pricing
- free
- Capabilities
- cross-framework inference, hardware acceleration, model optimization, operator pruning, quantization
- Integrations
- Android NNAPI, Apple Core ML, XNNPACK, DirectML, TensorRT
- Use Cases
- mobile model deployment, edge inference pipelines, cross-platform ML serving, real-time computer vision
- API Available
- No
- SDK Languages
- python, c++, c#, java, javascript
- Deployment
- edge, mobile, embedded, cloud
- Rate Limits
- No limits — fully on-device
- Data Privacy
- On-device processing; no cloud dependency required
- Tags
- edge, onnx, mobile, inference, cross-platform, iot
- Added
- 2026-03-17
- Completeness
- 100%
Index Score
70.4Adoption
82
Quality
88
Freshness
82
Citations
80
Engagement
0