QwQ-32B
by Alibaba / Qwen Team · free · Last verified 2026-03-17
QwQ-32B is a 32 billion parameter language model from Alibaba, specifically optimized for complex reasoning tasks. It utilizes a deep chain-of-thought methodology to excel at mathematical, scientific, and logical problems, achieving performance comparable to much larger models and showcasing high parameter efficiency.
https://huggingface.co/Qwen/QwQ-32B-Preview ↗B
B—Above Average
Adoption: BQuality: A+Freshness: ACitations: B+Engagement: F
Specifications
- License
- Apache 2.0
- Pricing
- free
- Capabilities
- deep-chain-of-thought-reasoning, advanced-mathematical-reasoning, multi-step-logical-inference, scientific-problem-solving, complex-code-analysis, symbolic-reasoning, problem-decomposition
- Integrations
- Use Cases
- [object Object], [object Object], [object Object], [object Object], [object Object]
- API Available
- Yes
- Parameters
- ~32B
- Context Window
- 32K
- Modalities
- text
- Training Cutoff
- 2024
- Tags
- reasoning, qwen, alibaba, chain-of-thought, math, large-language-model, open-source, parameter-efficient, scientific-reasoning, logical-reasoning
- Added
- 2026-03-17
- Completeness
- 0.85%
Index Score
62Adoption
65
Quality
90
Freshness
82
Citations
72
Engagement
0