Experimental Qwen variant—lean, fast, and a bit mysterious
7.4K
QwQ-32B is a 32-billion-parameter large language model designed to deliver high-level reasoning and intelligence. It achieves performance comparable to DeepSeek R1, a 671-billion-parameter model (with 37 billion activated), highlighting the efficiency of well-optimized foundation models trained on extensive world knowledge. The model incorporates agent-like capabilities, allowing it to perform critical reasoning, utilize tools, and adapt its behavior based on real-time environmental feedback. These features enable QwQ-32B to handle complex tasks with deep thinking and dynamic decision-making.
QwQ-32B is designed for tasks requiring advanced reasoning and problem-solving abilities.
| Attribute | Details |
|---|---|
| Provider | Alibaba Cloud |
| Architecture | qwen2 |
| Cutoff date | - |
| Languages | +29 |
| Tool calling | ✅ |
| Input modalities | Text |
| Output modalities | Text |
| License | Apache 2.0 |
| Model variant | Parameters | Quantization | Context window | VRAM¹ | Size |
|---|---|---|---|---|---|
ai/qwq:latestai/qwq:32B-Q4_K_M | 32B | IQ2_XXS/Q4_K_M | 41K tokens | 19.72 GiB | 18.48 GB |
ai/qwq:32B-Q4_0 | 32B | Q4_0 | 41K tokens | 18.60 GiB | 17.35 GB |
ai/qwq:32B-Q4_K_M | 32B | IQ2_XXS/Q4_K_M | 41K tokens | 19.72 GiB | 18.48 GB |
ai/qwq:32B-F16 | 32B | F16 | 41K tokens | 61.23 GiB | 61.03 GB |
¹: VRAM estimated based on model characteristics.
latest→32B-Q4_K_M
First, pull the model:
docker model pull ai/qwq
Then run the model:
docker model run ai/qwq
For more information on Docker Model Runner, explore the documentation.
Content type
Model
Digest
sha256:cd1dc0042…
Size
17.4 GB
Last updated
about 1 year ago
docker model pull ai/qwq:32B-Q4_0Pulls:
142
Last week