ai/qwq

Verified Publisher

By Docker

Updated about 1 year ago

Experimental Qwen variant—lean, fast, and a bit mysterious

Model
3

7.4K

ai/qwq repository overview

QwQ

logo

Description

QwQ-32B is a 32-billion-parameter large language model designed to deliver high-level reasoning and intelligence. It achieves performance comparable to DeepSeek R1, a 671-billion-parameter model (with 37 billion activated), highlighting the efficiency of well-optimized foundation models trained on extensive world knowledge. The model incorporates agent-like capabilities, allowing it to perform critical reasoning, utilize tools, and adapt its behavior based on real-time environmental feedback. These features enable QwQ-32B to handle complex tasks with deep thinking and dynamic decision-making.

Intended uses

QwQ-32B is designed for tasks requiring advanced reasoning and problem-solving abilities.

  • Mathematical problem solving: Excels in complex mathematical computations and proofs.
  • Code generation and debugging: Assists in writing and troubleshooting code across various programming languages.
  • General problem-solving: Provides insightful solutions to diverse challenges requiring logical reasoning.

Characteristics

AttributeDetails
ProviderAlibaba Cloud
Architectureqwen2
Cutoff date-
Languages+29
Tool calling
Input modalitiesText
Output modalitiesText
LicenseApache 2.0

Available model variants

Model variantParametersQuantizationContext windowVRAM¹Size
ai/qwq:latest

ai/qwq:32B-Q4_K_M
32BIQ2_XXS/Q4_K_M41K tokens19.72 GiB18.48 GB
ai/qwq:32B-Q4_032BQ4_041K tokens18.60 GiB17.35 GB
ai/qwq:32B-Q4_K_M32BIQ2_XXS/Q4_K_M41K tokens19.72 GiB18.48 GB
ai/qwq:32B-F1632BF1641K tokens61.23 GiB61.03 GB

¹: VRAM estimated based on model characteristics.

latest32B-Q4_K_M

Use this AI model with Docker Model Runner

First, pull the model:

docker model pull ai/qwq

Then run the model:

docker model run ai/qwq

For more information on Docker Model Runner, explore the documentation.

Considerations

  • Language mixing and code-switching: The model may unexpectedly switch languages, affecting response clarity.
  • Recursive reasoning loops: Potential for circular reasoning patterns leading to lengthy, inconclusive responses. Use Temperature=0.6, TopP=0.95, MinP=0 to avoid this and use TopK between 20 and 40 to filter out rare token occurrences while maintaining the diversity of the generated output.
  • Performance limitations: While excelling in math and coding, it may underperform in common sense reasoning and nuanced language understanding.

Tag summary

Content type

Model

Digest

sha256:cd1dc0042

Size

17.4 GB

Last updated

about 1 year ago

docker model pull ai/qwq:32B-Q4_0

This week's pulls

Pulls:

142

Last week