ai/granite-4.0-micro

Verified Publisher

By Docker

Updated 7 months ago

3B long-context instruct model with RL alignment, IF, tool use, and enterprise optimization.

Model
0

3.7K

ai/granite-4.0-micro repository overview

Granite-4.0-Micro

logo

Description

Granite-4.0-Micro is a 3B parameter long-context instruct model finetuned from Granite-4.0-Micro-Base using a combination of open source instruction datasets with permissive license and internally collected synthetic datasets. This model is developed using a diverse set of techniques with a structured chat format, including supervised finetuning, model alignment using reinforcement learning, and model merging. Granite 4.0 instruct models feature improved instruction following (IF) and tool-calling capabilities, making them more effective in enterprise applications.

Characteristics

AttributeDetails
ProviderGranite Team, IBM
Architecturegranitehybrid
Cutoff dateNot disclosed
LanguagesEnglish, German, Spanish, French, Japanese, Portuguese, Arabic, Czech, Italian, Korean, Dutch, Chinese (extensible via finetuning)
Tool calling
Input modalitiesText
Output modalitiesText
LicenseApache 2.0

Available model variants

Model variantParametersQuantizationContext windowVRAM¹Size
ai/granite-4.0-micro:3B

ai/granite-4.0-micro:3B-Q4_K_M

ai/granite-4.0-micro:latest
3.2BMOSTLY_Q4_K_M1M tokens2.32 GiB1.81 GB

¹: VRAM estimated based on model characteristics.

latest3B

Use this AI model with Docker Model Runner

docker model run ai/granite-4.0-micro

Considerations

  • Optimized for instruction following, tool/function calling, and long-context (up to 128K tokens) scenarios.
  • Strong generalist capabilities: summarization, classification, extraction, QA/RAG, coding, function-calling, and multilingual dialogue.
  • Multilingual: best performance in English; a few-shot approach or light finetuning can help close gaps for other languages.
  • Safety & reliability: despite alignment, the model can still produce inaccurate or biased outputs—apply domain-specific evaluation and guardrails.
  • Infrastructure note: trained on NVIDIA GB200 NVL72 at CoreWeave; use acceleration libraries (e.g., accelerate, optimized attention/KV cache settings) for efficient inference.

Benchmark performance

CategoryMetricGranite-4.0-Micro
General Tasks
MMLU (5-shot)65.98
MMLU-Pro (5-shot, CoT)44.50
BBH (3-shot, CoT)72.48
AGI EVAL (0-shot, CoT)64.29
GPQA (0-shot, CoT)30.14
Alignment Tasks
AlpacaEval 2.029.49
IFEval (Instruct, Strict)85.50
IFEval (Prompt, Strict)79.12
IFEval (Average)82.31
ArenaHard25.84
Math Tasks
GSM8K (8-shot)85.45
GSM8K Symbolic (8-shot)79.82
Minerva Math (0-shot, CoT)62.06
DeepMind Math (0-shot, CoT)44.56
Code Tasks
HumanEval (pass@1)80.00
HumanEval+ (pass@1)72.00
MBPP (pass@1)72.00
MBPP+ (pass@1)64.00
CRUXEval-O (pass@1)41.50
BigCodeBench (pass@1)39.21
Tool Calling Tasks
BFCL v359.98
Multilingual Tasks
MULTIPLE (pass@1)49.21
MMMLU (5-shot)55.14
INCLUDE (5-shot)51.62
MGSM (8-shot)28.56
Safety
SALAD-Bench97.06
AttaQ86.05

Tag summary

Content type

Model

Digest

sha256:5ce5a9ac1

Size

1.8 GB

Last updated

7 months ago

docker model pull ai/granite-4.0-micro

This week's pulls

Pulls:

64

Last week