Granite Embedding Multilingual is a 278 million parameter, encoder‑only XLM‑RoBERTa‑style
6.5K
Granite Embedding Multilingual is a 278 million parameter, encoder‑only XLM‑RoBERTa‑style dense biencoder model created by IBM, engineered to produce high‑quality multilingual text embeddings (768‑dimensional). It’s optimized for semantic similarity, retrieval, and search in 12 major languages, released under the Apache 2.0 license.
Designed for fixed-length vector generation suitable for multilingual search and retrieval tasks:
| Attribute | Details |
|---|---|
| Provider | IBM (Granite Embedding Team) |
| Architecture | Encoder‑only transformer, XLM‑RoBERTa‑like bi‑encoder |
| Cutoff date | Released December 18, 2024:contentReference |
| Languages | Multilingual: English, German, Spanish, French, Japanese, Portuguese, Arabic, Czech, Italian, Korean, Dutch, Chinese |
| Tool calling | No (not for tool‑calling; it's an embedding model) |
| Input modalities | Text (up to 512 tokens per input) |
| Output modalities | Fixed-length embedding vectors (768 dimensions) |
| License | Apache 2.0 |
| Model variant | Parameters | Quantization | Context window | VRAM¹ | Size |
|---|---|---|---|---|---|
ai/granite-embedding-multilingual:latestai/granite-embedding-multilingual:278M-F16 | 278M | MOSTLY_F16 | 512 tokens | 0.19 GiB | 530.18 MB |
ai/granite-embedding-multilingual:278M-F16 | 278M | MOSTLY_F16 | 512 tokens | 0.19 GiB | 530.18 MB |
¹: VRAM estimated based on model characteristics.
latest→278M-F16
First, pull the model:
docker model pull {model_name}
Then run the model:
docker model run {model_name}
Content type
Model
Digest
sha256:69c965354…
Size
536.7 MB
Last updated
9 months ago
docker model pull ai/granite-embedding-multilingualPulls:
85
Last week