Embedding Gemma is a state-of-the-art text embedding model from Google DeepMind
3.5K
Embedding Gemma is a state-of-the-art text embedding model from Google DeepMind, designed to create high-quality vector representations of text. Built on the Gemma architecture, this model converts text into dense vector embeddings that capture semantic meaning, making it ideal for retrieval-augmented generation (RAG), semantic search, and similarity tasks. With open weights and efficient design, Embedding Gemma provides a powerful foundation for embedding-based applications.
Embedding Gemma is designed for applications requiring high-quality text embeddings:
| Attribute | Details |
|---|---|
| Provider | Google DeepMind |
| Architecture | Gemma Embedding |
| Cutoff date | - |
| Languages | English |
| Tool calling | ❌ |
| Input modalities | Text |
| Output modalities | Embedding vectors |
| License | Gemma Terms |
First, pull the model:
docker model pull ai/embeddinggemma-vllm
To generate embeddings using the API:
curl --location 'http://localhost:12434/engines/llama.cpp/v1/embeddings' \
--header 'Content-Type: application/json' \
--data '{
"model": "ai/embeddinggemma-vllm",
"input": "Your text to embed here"
}'
For more information on Docker Model Runner, explore the documentation.
| Task Category | Embedding Gemma |
|---|---|
| Retrieval | 54.87 |
| STS | 78.53 |
| Classification | 73.26 |
| Clustering | 44.72 |
| Pair Classification | 85.94 |
| Reranking | 59.36 |
Content type
Model
Digest
sha256:907318686…
Size
1.2 GB
Last updated
6 months ago
docker model pull ai/embeddinggemma-vllm:300MPulls:
108
Last week