ai/granite-embedding-multilingual

Verified Publisher

By Docker

Updated 9 months ago

Granite Embedding Multilingual is a 278 million parameter, encoder‑only XLM‑RoBERTa‑style

Model
3

6.5K

ai/granite-embedding-multilingual repository overview

Granite Embedding Multilingual

logo

Granite Embedding Multilingual is a 278 million parameter, encoder‑only XLM‑RoBERTa‑style dense biencoder model created by IBM, engineered to produce high‑quality multilingual text embeddings (768‑dimensional). It’s optimized for semantic similarity, retrieval, and search in 12 major languages, released under the Apache 2.0 license.

Intended uses

Designed for fixed-length vector generation suitable for multilingual search and retrieval tasks:

  • Semantic similarity and retrieval: Compute vector representations for efficient similarity comparisons in multilingual contexts
  • Cross-lingual information retrieval (RAG, search): Works across 12 languages for tasks like clustering or search
  • Enterprise-grade deployment: Built with ethically sourced, enterprise‑friendly datasets and transparent processes

Characteristics

AttributeDetails
ProviderIBM (Granite Embedding Team)
ArchitectureEncoder‑only transformer, XLM‑RoBERTa‑like bi‑encoder
Cutoff dateReleased December 18, 2024:contentReference
LanguagesMultilingual: English, German, Spanish, French, Japanese, Portuguese, Arabic, Czech, Italian, Korean, Dutch, Chinese
Tool callingNo (not for tool‑calling; it's an embedding model)
Input modalitiesText (up to 512 tokens per input)
Output modalitiesFixed-length embedding vectors (768 dimensions)
LicenseApache 2.0

Available model variants

Model variantParametersQuantizationContext windowVRAM¹Size
ai/granite-embedding-multilingual:latest

ai/granite-embedding-multilingual:278M-F16
278MMOSTLY_F16512 tokens0.19 GiB530.18 MB
ai/granite-embedding-multilingual:278M-F16278MMOSTLY_F16512 tokens0.19 GiB530.18 MB

¹: VRAM estimated based on model characteristics.

latest278M-F16

Use this AI model with Docker Model Runner

First, pull the model:

docker model pull {model_name}

Then run the model:

docker model run {model_name}

Considerations

  • Context is limited to 512 tokens—longer inputs need truncation or chunking.
  • Performance is strong on multilingual.
  • Built following IBM’s AI ethics guidelines, with transparent dataset governance and licensing.

Tag summary

Content type

Model

Digest

sha256:69c965354

Size

536.7 MB

Last updated

9 months ago

docker model pull ai/granite-embedding-multilingual

This week's pulls

Pulls:

85

Last week