3B long-context instruct model with RL alignment, IF, tool use, and enterprise optimization.
3.7K
Granite-4.0-Micro is a 3B parameter long-context instruct model finetuned from Granite-4.0-Micro-Base using a combination of open source instruction datasets with permissive license and internally collected synthetic datasets. This model is developed using a diverse set of techniques with a structured chat format, including supervised finetuning, model alignment using reinforcement learning, and model merging. Granite 4.0 instruct models feature improved instruction following (IF) and tool-calling capabilities, making them more effective in enterprise applications.
| Attribute | Details |
|---|---|
| Provider | Granite Team, IBM |
| Architecture | granitehybrid |
| Cutoff date | Not disclosed |
| Languages | English, German, Spanish, French, Japanese, Portuguese, Arabic, Czech, Italian, Korean, Dutch, Chinese (extensible via finetuning) |
| Tool calling | ✅ |
| Input modalities | Text |
| Output modalities | Text |
| License | Apache 2.0 |
| Model variant | Parameters | Quantization | Context window | VRAM¹ | Size |
|---|---|---|---|---|---|
ai/granite-4.0-micro:3Bai/granite-4.0-micro:3B-Q4_K_Mai/granite-4.0-micro:latest | 3.2B | MOSTLY_Q4_K_M | 1M tokens | 2.32 GiB | 1.81 GB |
¹: VRAM estimated based on model characteristics.
latest→3B
docker model run ai/granite-4.0-micro
| Category | Metric | Granite-4.0-Micro |
|---|---|---|
| General Tasks | ||
| MMLU (5-shot) | 65.98 | |
| MMLU-Pro (5-shot, CoT) | 44.50 | |
| BBH (3-shot, CoT) | 72.48 | |
| AGI EVAL (0-shot, CoT) | 64.29 | |
| GPQA (0-shot, CoT) | 30.14 | |
| Alignment Tasks | ||
| AlpacaEval 2.0 | 29.49 | |
| IFEval (Instruct, Strict) | 85.50 | |
| IFEval (Prompt, Strict) | 79.12 | |
| IFEval (Average) | 82.31 | |
| ArenaHard | 25.84 | |
| Math Tasks | ||
| GSM8K (8-shot) | 85.45 | |
| GSM8K Symbolic (8-shot) | 79.82 | |
| Minerva Math (0-shot, CoT) | 62.06 | |
| DeepMind Math (0-shot, CoT) | 44.56 | |
| Code Tasks | ||
| HumanEval (pass@1) | 80.00 | |
| HumanEval+ (pass@1) | 72.00 | |
| MBPP (pass@1) | 72.00 | |
| MBPP+ (pass@1) | 64.00 | |
| CRUXEval-O (pass@1) | 41.50 | |
| BigCodeBench (pass@1) | 39.21 | |
| Tool Calling Tasks | ||
| BFCL v3 | 59.98 | |
| Multilingual Tasks | ||
| MULTIPLE (pass@1) | 49.21 | |
| MMMLU (5-shot) | 55.14 | |
| INCLUDE (5-shot) | 51.62 | |
| MGSM (8-shot) | 28.56 | |
| Safety | ||
| SALAD-Bench | 97.06 | |
| AttaQ | 86.05 |
Content type
Model
Digest
sha256:5ce5a9ac1…
Size
1.8 GB
Last updated
7 months ago
docker model pull ai/granite-4.0-microPulls:
64
Last week