ai/granite-4.0-nano

Verified Publisher

By Docker

Updated 6 months ago

Granite-4.0-nano: lightweight instruct model trained via SFT, RL, and merging on diverse data.

Model
0

9.9K

ai/granite-4.0-nano repository overview

Granite 4.0 Nano

logo

Description

Granite-4.0-350M is a lightweight instruct model finetuned from Granite-4.0-350M-Base using a combination of open source instruction datasets with permissive license and internally collected synthetic datasets. This model is developed using a diverse set of techniques including supervised finetuning, reinforcement learning, and model merging.

Characteristics

AttributeDetails
ProviderGranite Team, IBM
Architecturegranitehybrid
Cutoff dateNot disclosed
LanguagesEnglish, German, Spanish, French, Japanese, Portuguese, Arabic, Czech, Italian, Korean, Dutch, Chinese (extensible via finetuning)
Tool calling
Input modalitiesText
Output modalitiesText
LicenseApache 2.0

Intended use

Intended use: Granite 4.0 Nano instruct models feature strong instruction following capabilities bringing advanced AI capabilities within reach for on-device deployments and research use cases. Additionally, their compact size makes them well-suited for fine-tuning on specialized domains without requiring massive compute resources.

Available model variants

Model variantParametersQuantizationContext windowVRAM¹Size
ai/granite-4.0-nano:1B

ai/granite-4.0-nano:1B-BF16

ai/granite-4.0-nano:latest
1BMOSTLY_BF16131K tokens3.89 GiB3.04 GB
ai/granite-4.0-nano:350M-BF16350MMOSTLY_BF1633K tokens1.29 GiB672.22 MB

¹: VRAM estimated based on model characteristics.

latest1B

Use this AI model with Docker Model Runner

docker model run ai/granite-4.0-nano

Considerations

  • Optimized for instruction following, tool/function calling, and long-context (up to 128K tokens) scenarios.
  • Strong generalist capabilities: summarization, classification, extraction, QA/RAG, coding, function-calling, and multilingual dialogue.
  • Multilingual: best performance in English; a few-shot approach or light finetuning can help close gaps for other languages.
  • Safety & reliability: despite alignment, the model can still produce inaccurate or biased outputs—apply domain-specific evaluation and guardrails.
  • Infrastructure note: trained on NVIDIA GB200 NVL72 at CoreWeave; use acceleration libraries (e.g., accelerate, optimized attention/KV cache settings) for efficient inference.

Benchmark performance

BenchmarksMetric350M DenseH 350M Dense1B DenseH 1B Dense
General Tasks
MMLU5-shot35.0136.2159.3959.74
MMLU-Pro5-shot, CoT12.1314.3834.0232.86
BBH3-shot, CoT33.0733.2860.3759.68
AGI EVAL0-shot, CoT26.2229.6149.2252.44
GPQA0-shot, CoT24.1126.1229.9129.69
Alignment Tasks
IFEvalInstruct, Strict61.6367.6380.8282.37
IFEvalPrompt, Strict49.1755.6473.9474.68
IFEvalAverage55.4061.6377.3878.53
Math Tasks
GSM8K8-shot30.7139.2776.3569.83
GSM Symbolic8-shot26.7633.7072.3065.72
Minerva Math0-shot, CoT13.045.7645.2849.40
DeepMind Math0-shot, CoT8.456.2034.0034.98
Code Tasks
HumanEvalpass@139.0038.0074.0073.00
HumanEval+pass@137.0035.0069.0068.00
MBPPpass@148.0049.0065.0069.00
MBPP+pass@138.0044.0057.0060.00
CRUXEval-Opass@123.7525.5033.1336.00
BigCodeBenchpass@111.1411.2330.1829.12
Tool Calling Tasks
BFCL v339.3243.3254.8250.21
Multilingual Tasks
MULTIPLEpass@115.9914.3132.2436.11
MMMLU5-shot28.2327.9545.0049.43
INCLUDE5-shot27.7427.0942.1243.35
MGSM8-shot14.7216.1637.8427.52
Safety
SALAD-Bench97.1296.5593.4496.40
AttaQ82.5381.7685.2682.85

Tag summary

Content type

Model

Digest

sha256:34ae9a653

Size

3 GB

Last updated

6 months ago

docker model pull ai/granite-4.0-nano

This week's pulls

Pulls:

147

Last week