DeepSeek-Coder-V2
by DeepSeek · freemium · Last verified 2026-03-17
DeepSeek-Coder-V2 is a powerful open-source Mixture-of-Experts (MoE) model specialized in code. It supports 338 programming languages and features advanced fill-in-the-middle capabilities, offering performance comparable to top-tier proprietary models like GPT-4 Turbo at a significantly lower inference cost.
https://www.deepseek.com ↗B
B—Above Average
Adoption: B+Quality: AFreshness: BCitations: B+Engagement: F
Specifications
- License
- DeepSeek License
- Pricing
- freemium
- Capabilities
- Code generation and completion, Fill-in-the-middle (FIM) code insertion, Multi-language code translation, Automated unit test generation, Code debugging and error fixing, Algorithm implementation from descriptions, Code refactoring and optimization, Mathematical reasoning, API usage and documentation generation
- Integrations
- Hugging Face Transformers, vLLM, Ollama, LangChain, LlamaIndex, VS Code Extensions, JetBrains IDE Plugins
- Use Cases
- [object Object], [object Object], [object Object], [object Object], [object Object]
- API Available
- Yes
- Parameters
- 236B (21B active)
- Context Window
- 128K tokens
- Modalities
- text
- Training Cutoff
- Mid 2024
- Tags
- code-generation, open-source, moe, mixture-of-experts, fill-in-middle, code-llm, software-development, multi-language, self-hosting, deepseek, transformer
- Added
- 2026-03-17
- Completeness
- 1%
Index Score
65.4Adoption
74
Quality
84
Freshness
65
Citations
76
Engagement
0