Skip to main content
ScriptAI Infrastructurev1.2

Edge Model Optimization

by Community · open-source · Last verified 2026-03-17

Optimizes PyTorch or TensorFlow models for edge deployment by applying INT8/FP16 quantization, ONNX export, and TFLite conversion with platform-specific tuning for ARM/NPU targets. Benchmarks latency, memory, and accuracy trade-offs across optimization strategies and generates a deployment report.

https://github.com/NVIDIA/TensorRT
C+
C+Average
Adoption: BQuality: AFreshness: ACitations: C+Engagement: F

Specifications

License
Apache-2.0
Pricing
open-source
Capabilities
int8-quantization, onnx-export, tflite-conversion, benchmark-report
Integrations
onnxruntime, tensorrt, tflite, torchscript
Use Cases
mobile-ml, iot-inference, raspberry-pi-deployment
API Available
No
Language
python
Dependencies
torch, onnx, onnxruntime, tensorflow-lite-runtime, numpy
Environment
Python 3.10+
Est. Runtime
5-30 minutes per model
Tags
edge-deployment, onnx, quantization, tflite, model-compression
Added
2026-03-17
Completeness
100%

Index Score

58.7
Adoption
68
Quality
85
Freshness
82
Citations
58
Engagement
0

Explore the full AI ecosystem on Agents as a Service