Skip to main content
IntegrationAI Infrastructurev0.9.x

LlamaFile + Local Execution

by Mozilla · open-source · Last verified 2026-03-17

LlamaFile by Mozilla and Justine Tunney bundles a complete LLM with its runtime into a single self-contained executable that runs on Linux, macOS, Windows, FreeBSD, NetBSD, and OpenBSD without any installation. It embeds a compressed GGUF model and a llama.cpp backend into a polyglot binary (ZIP + ELF/Mach-O), serving an OpenAI-compatible HTTP API on localhost at startup.

https://github.com/Mozilla-Ocho/llamafile
C+
C+Average
Adoption: C+Quality: AFreshness: ACitations: CEngagement: F

Specifications

License
Apache-2.0
Pricing
open-source
Capabilities
single-binary-deployment, cross-platform, openai-compatible-api, no-install, gpu-acceleration
Integrations
llama-cpp, local-execution
Use Cases
air-gapped-inference, developer-local-testing, edge-deployment, portable-llm-distribution
API Available
Yes
Tags
local-inference, single-binary, portable, llama-cpp, edge
Added
2026-03-17
Completeness
100%

Index Score

52
Adoption
58
Quality
84
Freshness
88
Citations
48
Engagement
0

Put AI to work for your business

Deploy this integration alongside autonomous AaaS agents that handle tasks end-to-end — no babysitting required.

Explore the full AI ecosystem on Agents as a Service