Ollama + Docker
by Ollama · open-source · Last verified 2026-03-17
Ollama's official Docker image packages the Ollama runtime for containerized local LLM inference, enabling teams to run quantized GGUF models on CPU or GPU inside Docker Compose stacks or Kubernetes pods. The integration supports GPU passthrough via NVIDIA Container Toolkit and provides an OpenAI-compatible HTTP API for drop-in compatibility with existing tooling.
https://ollama.com ↗B
B—Above Average
Adoption: AQuality: AFreshness: A+Citations: B+Engagement: F
Specifications
- License
- MIT
- Pricing
- open-source
- Capabilities
- containerized-inference, gpu-passthrough, openai-compatible-api, model-library, cpu-fallback
- Integrations
- docker, kubernetes, nvidia-container-toolkit
- Use Cases
- local-llm-development, offline-inference, docker-compose-llm-stack, ci-testing-with-llms
- API Available
- Yes
- Tags
- local-inference, docker, self-hosted, gguf, developer-tools
- Added
- 2026-03-17
- Completeness
- 100%
Index Score
67.5Adoption
82
Quality
86
Freshness
92
Citations
70
Engagement
0
Put AI to work for your business
Deploy this integration alongside autonomous AaaS agents that handle tasks end-to-end — no babysitting required.