State-of-the-art text embedding via the Gemini API

State-of-the-art text embedding via the Gemini API

[Image created by Google with Gemini 2.0 Flash native image generation]

Right now, we’re making a brand new experimental Gemini Embedding textual content mannequin (gemini-embedding-exp-03-07)1 obtainable within the Gemini API.

Educated on the Gemini mannequin itself, this embedding mannequin has inherited Gemini’s understanding of language and nuanced context making it relevant for a variety of makes use of. This new embedding mannequin surpasses our earlier state-of-the-art mannequin (text-embedding-004), achieves the highest rank on the Huge Textual content Embedding Benchmark (MTEB) Multilingual leaderboard, and comes with new options like longer enter token size!


Our most succesful textual content embedding mannequin but

We have educated our mannequin to be remarkably normal, delivering distinctive efficiency throughout various domains, together with finance, science, authorized, search, and extra. It really works successfully out-of-the-box, eliminating the necessity for intensive fine-tuning for particular duties.

The MTEB (Multilingual) leaderboard ranks textual content embedding fashions throughout various duties reminiscent of retrieval and classification to offer a complete benchmark for mannequin comparability. Our Gemini Embedding mannequin achieves a imply (job) rating of 68.32–a margin of +5.81 over the following competing mannequin.

Our new Gemini textual content embedding mannequin (gemini-embedding-exp-03-07) achieves excessive scores on the MTEB (Multilingual) leaderboard (proper click on to open picture in new tab).

Why embeddings?

From constructing clever retrieval augmented era (RAG) and advice techniques to textual content classification, the power for LLMs to know the that means behind textual content is essential. Embeddings are sometimes essential for constructing extra environment friendly techniques, decreasing price and latency whereas additionally usually offering higher outcomes than key phrase matching techniques. Embeddings seize semantic that means and context by way of numerical representations of knowledge. Knowledge with comparable semantic that means have embeddings which might be nearer collectively. Embeddings allow a variety of purposes, together with:

  • Environment friendly Retrieval: Discover related paperwork inside giant databases, like authorized doc retrieval or enterprise search, by evaluating the embeddings of queries and paperwork.
  • Retrieval-Augmented Era (RAG): Improve the standard and relevance of generated textual content by retrieving and incorporating contextually related info into the context of a mannequin.
  • Clustering and Categorization: Group comparable texts collectively, figuring out traits and matters inside your knowledge.
  • Classification: Mechanically categorize textual content based mostly on its content material, reminiscent of sentiment evaluation or spam detection.
  • Textual content Similarity: Determine duplicate content material, enabling duties like net web page deduplication or plagiarism detection.

You’ll be able to be taught extra about embeddings and customary AI use instances within the Gemini API docs.


Get began with Gemini Embedding

Builders can now entry our new, experimental Gemini Embeddings mannequin by way of the Gemini API. It’s suitable with the prevailing embed_content endpoint.

from google import genai

shopper = genai.Shopper(api_key="GEMINI_API_KEY")

end result = shopper.fashions.embed_content(
        mannequin="gemini-embedding-exp-03-07",
        contents="How does alphafold work?",
        )

print(end result.embeddings)

Along with improved high quality throughout all dimensions, Gemini Embedding additionally options:

  • Enter token restrict of 8K tokens. We’ve improved our context size from earlier fashions permitting you to embed giant chunks of textual content, code, or different knowledge.
  • Output dimensions of 3K dimensions. Excessive-dimensional embeddings with virtually 4x extra tokens over earlier embedding fashions.
  • Matryoshka Illustration Studying (MRL): MRL means that you can truncate the unique 3K dimensions to scale down to satisfy your required storage price.
  • Expanded language assist. We’ve doubled the variety of languages supported to over 100.
  • Unified mannequin. This mannequin surpasses the standard of our earlier task-specific multilingual, english-only, and code particular fashions.

Whereas at the moment in an experimental section with restricted capability, this launch offers you an early alternative to discover Gemini Embedding capabilities. As with all experimental fashions, it is topic to alter, and we’re working in direction of a steady, usually obtainable launch within the months to come back. We’d love to listen to your suggestions on the embeddings feedback form.


1 On Vertex AI, the identical mannequin is served by way of the endpoint “text-embedding-large-exp-03-07.” For normal availability, naming can be constant.

Leave a Reply