ai/qwen3.6

Verified Publisher

By Docker

•Updated 7 days ago

Multimodal AI model with 35B MoE architecture for coding agents, reasoning, and vision tasks

Model
1

10K+

ai/qwen3.6 repository overview

⁠Qwen3.6

Qwen Logo

Qwen3.6-35B-A3B is a cutting-edge multimodal AI model from Alibaba Cloud's Qwen team, delivering substantial upgrades in agentic coding, reasoning preservation, and real-world development workflows. Built on direct community feedback following the Qwen3.5 series, this model prioritizes stability and utility, offering developers an intuitive and productive coding experience with enhanced frontend workflows and repository-level reasoning capabilities.

This release features a Mixture of Experts (MoE) architecture with 35 billion total parameters and 3 billion activated parameters, supporting both text and vision inputs. With a native context length of 262,144 tokens (extensible to over 1 million), Qwen3.6 excels at handling complex, multi-turn interactions while maintaining efficiency. The model demonstrates exceptional performance on coding agent benchmarks like SWE-bench Verified, Terminal-Bench, and web development tasks, making it particularly valuable for software engineering, code generation, and multimodal applications.

Key innovations include thinking preservation to retain reasoning context across iterations, improved tool calling with better nested object parsing, and enhanced support for developer roles. Whether you're building coding assistants, multimodal applications, or agentic systems, Qwen3.6 delivers state-of-the-art performance with open-source accessibility.


⁠Characteristics

AttributeValue
ProviderAlibaba Cloud / Qwen Team
ArchitectureQwen3.5 MoE (Mixture of Experts)
Parameters35B total, 3B activated
Context Length262,144 tokens (native), extensible to 1,010,000 tokens
LanguagesMultilingual (English, Chinese, and others)
Input modalitiesText, Image
Output modalitiesText
LicenseApache 2.0

⁠Using this model with Docker Model Runner

docker model run qwen3.6

For more information, check out the Docker Model Runner docs⁠.

⁠Benchmarks

Benchmark Results

⁠Coding Agent Benchmarks
BenchmarkQwen3.5-27BGemma4-31BQwen3.5-35BA3BGemma4-26BA4BQwen3.6-35BA3B
SWE-bench Verified75.052.070.017.473.4
SWE-bench Multilingual69.351.760.317.367.2
SWE-bench Pro51.235.744.613.849.5
Terminal-Bench 2.041.642.940.534.251.5
Claw-Eval (Avg)64.348.565.458.868.7
Claw-Eval (Pass³)46.225.051.028.050.0
SkillsBench (Avg5)27.223.64.412.328.7
QwenClawBench52.241.747.738.752.6
NL2Repo27.315.520.511.629.4
QwenWebBench1068119797811781397
⁠General Agent Benchmarks
BenchmarkQwen3.5-27BGemma4-31BQwen3.5-35BA3BGemma4-26BA4BQwen3.6-35BA3B
TAU3-Bench68.467.568.959.067.2
VITA-Bench41.843.029.136.935.6
DeepPlanning22.624.022.816.225.9
Tool Decathlon31.521.228.712.026.9
MCPMark36.318.127.014.237.0
MCP-Atlas68.457.262.450.062.8
WideSearch66.435.259.138.360.1
⁠Knowledge Benchmarks
BenchmarkQwen3.5-27BGemma4-31BQwen3.5-35BA3BGemma4-26BA4BQwen3.6-35BA3B
MMLU-Pro86.185.285.382.685.2
MMLU-Redux93.293.793.392.793.3
SuperGPQA65.665.763.461.464.7
C-Eval90.582.690.282.590.0
⁠STEM & Reasoning Benchmarks
BenchmarkQwen3.5-27BGemma4-31BQwen3.5-35BA3BGemma4-26BA4BQwen3.6-35BA3B
GPQA85.584.384.282.386.0
HLE24.319.522.48.721.4
LiveCodeBench v680.780.074.677.180.4
HMMT Feb 2592.088.789.091.790.7
HMMT Nov 2589.887.589.287.589.1
HMMT Feb 2684.377.278.779.083.6
IMOAnswerBench79.974.576.874.378.9
AIME2692.689.291.088.392.7
⁠Vision Language Benchmarks
BenchmarkQwen3.5-27BClaude-Sonnet-4.5Gemma4-31BGemma4-26BA4BQwen3.5-35BA3BQwen3.6-35BA3B
MMMU82.379.680.478.481.481.7
MMMU-Pro75.068.476.973.875.175.3
MathVista (mini)87.879.887.586.387.187.6
MathVision32.431.529.724.832.033.3
AI2D98.096.298.097.297.898.1

⁠Considerations

  • Hardware Requirements: Due to the MoE architecture with 35B total parameters, running this model efficiently requires adequate GPU memory. GGUF quantized versions are available to reduce memory requirements while maintaining performance.
  • Context Window: While the native context length is 262K tokens, optimal performance may vary at extreme lengths. For most applications, the native 262K context provides excellent performance.
  • Multimodal Capabilities: The model supports both text and image inputs. Vision capabilities require the appropriate multimodal projection files (mmproj files available in GGUF format).
  • Tool Calling: Qwen3.6 includes improved tool calling support with better handling of nested objects. Follow the documented chat template format for optimal tool use.
  • Developer Role Support: This model includes enhanced support for developer roles, making it particularly effective for coding assistants and agentic workflows.
  • Thinking Preservation: When using the model for iterative development, enable thinking preservation to maintain reasoning context across multiple turns.
⁠Generated by

This model card was automatically generated using cagent-action⁠. Want to learn more about Docker Model Runner? Check out the project repository: https://github.com/docker/model-runner⁠.

Tag summary

Content type

Model

Digest

sha256:8ca73612f…

Size

17.3 GB

Last updated

7 days ago

docker model pull ai/qwen3.6:27B

This week's pulls

Pulls:

10,969

Last week