LlamaFile + Local Execution
by Mozilla · open-source · Last verified 2026-03-17
LlamaFile by Mozilla and Justine Tunney bundles a complete LLM with its runtime into a single self-contained executable that runs on Linux, macOS, Windows, FreeBSD, NetBSD, and OpenBSD without any installation. It embeds a compressed GGUF model and a llama.cpp backend into a polyglot binary (ZIP + ELF/Mach-O), serving an OpenAI-compatible HTTP API on localhost at startup.
https://github.com/Mozilla-Ocho/llamafile ↗C+
C+—Average
Adoption: C+Quality: AFreshness: ACitations: CEngagement: F
Specifications
- License
- Apache-2.0
- Pricing
- open-source
- Capabilities
- single-binary-deployment, cross-platform, openai-compatible-api, no-install, gpu-acceleration
- Integrations
- llama-cpp, local-execution
- Use Cases
- air-gapped-inference, developer-local-testing, edge-deployment, portable-llm-distribution
- API Available
- Yes
- Tags
- local-inference, single-binary, portable, llama-cpp, edge
- Added
- 2026-03-17
- Completeness
- 100%
Index Score
52Adoption
58
Quality
84
Freshness
88
Citations
48
Engagement
0
Put AI to work for your business
Deploy this integration alongside autonomous AaaS agents that handle tasks end-to-end — no babysitting required.