Understanding Edges
RAG Edges
Retrieve Edge
What Is an Embedding Model?

What is an Embedding Model?


An embedding model converts text into a high-dimensional numeric vector. This vector captures the semantic meaning of the text, so that similar texts will have similar vectors—even if they use different words.

For example:

“What is your return policy?” and “How do I send an item back?” will generate vectors that are close in space.

These embedding vectors are essential for effective semantic search, because they allow the system to understand meaning, not just keywords.


How Vector Similarity Works?

When a user inputs a query, it is converted into a vector using the same embedding model as your documents. The system then computes the similarity between the query vector and all indexed vectors.

There are different similarity computation methods:

Cosine Similarity (default and most common): Measures the angle between two vectors. Values range from -1 to 1. A score close to 1 means high similarity.

Dot Product: Measures the projection of one vector onto another. Sensitive to both direction and magnitude.

Hamming Distance: Measures how many components differ between two binary vectors. Less common for text embeddings.

You can choose the similarity metric when you index the document collection.


What is a Token?

Tokens are the units of text processed by models. A token could be a word, part of a word, or even just punctuation.

“unbelievable” = 1 word = ~3–4 tokens

“Hi!” = ~2 tokens

Knowing token limits helps you select embedding models and chunking strategies appropriately. For example, if your model has a 512-token limit, ensure each chunk doesn’t exceed that.


How to Choose a Good Embedding Model

When selecting a model, consider:

Language support: Choose multilingual models if you handle data in different languages.

Task match: Some models are optimized for sentence similarity, others for retrieval or document search.

Token capacity: Higher token limits are better for longer inputs.

Embedding size: Larger vectors (e.g. 1024, 1536) capture more semantic information but may be slower.

Refer to the embedding model table below for comparison across providers.

ProviderModel NameVector DimensionsMax Text LengthLanguage SupportTasks
Sentence Transformerparaphrase-multilingual-mpnet-base-v2768512 tokensMultilingualSemantic search, clustering, sentence similarity
Sentence Transformerparaphrase-multilingual-MiniLM-L12-v2384256 tokensMultilingualParaphrase detection, semantic search
Sentence Transformerparaphrase-albert-small-v2768128 tokensEnglishParaphrase detection, semantic similarity
Sentence Transformerparaphrase-MiniLM-L3-v2384256 tokensEnglishParaphrase detection, semantic search
Sentence Transformermulti-qa-mpnet-base-dot-v1768512 tokensEnglishQuestion answering, information retrieval
Sentence Transformermulti-qa-distilbert-cos-v1768512 tokensEnglishQuestion answering, semantic search
Sentence Transformermulti-qa-MiniLM-L6-cos-v1384256 tokensEnglishQuestion answering, semantic search
Sentence Transformerdistiluse-base-multilingual-cased-v2512512 tokensMultilingualSemantic search, sentence similarity
Sentence Transformerdistiluse-base-multilingual-cased-v1512512 tokensMultilingualSemantic search, sentence similarity
Sentence Transformerall-mpnet-base-v2768512 tokensEnglishSemantic search, clustering, sentence similarity
Sentence Transformerall-distilroberta-v1768512 tokensEnglishSemantic search, sentence similarity
Sentence Transformerall-MiniLM-L6-v2384256 tokensEnglishSemantic search, sentence similarity
Sentence Transformerall-MiniLM-L12-v2384256 tokensEnglishSemantic search, sentence similarity
Hugging FaceBAAI/bge-m31024512 tokensMultilingualRetrieval-augmented generation, semantic search
Hugging FaceBAAI/llm-embedder1024512 tokensEnglishLanguage model embedding, retrieval tasks
Hugging FaceBAAI/bge-large-en-v1.51024512 tokensEnglishRetrieval-augmented generation, semantic search
Hugging FaceBAAI/bge-base-en-v1.5768512 tokensEnglishRetrieval-augmented generation, semantic search
Hugging FaceBAAI/bge-small-en-v1.5384512 tokensEnglishRetrieval-augmented generation, semantic search
Hugging FaceBAAI/bge-large-zh-v1.51024512 tokensChineseRetrieval-augmented generation, semantic search
Hugging FaceBAAI/bge-base-zh-v1.5768512 tokensChineseRetrieval-augmented generation, semantic search
Hugging FaceBAAI/bge-small-zh-v1.5384512 tokensChineseRetrieval-augmented generation, semantic search
Hugging FaceDMetaSoul/Dmeta-embedding-zh768512 tokensChineseSemantic search, sentence similarity
Hugging Faceshibing624/text2vec-base-chinese768512 tokensChineseSentence/document embeddings, semantic similarity
Hugging Facesentence-transformers/sentence-t5-large1024512 tokensMultilingualText generation, summarization, translation
Hugging Facesentence-transformers/mpnet768512 tokensMultilingualGeneral text embeddings, semantic search
Hugging Facejinaai/jina-colbert-v2768512 tokensMultilingualLate interaction retrieval, semantic search
Hugging Facejinaai/jina-embeddings-v310248192 tokensMultilingualLong-context retrieval, semantic search
Hugging Facejinaai/jina-embeddings-v2-base-zh768512 tokensChinese, EnglishBilingual embeddings, semantic search
Hugging Faceopenbmb/MiniCPM-Embedding1024512 tokensChinese, EnglishRetrieval tasks, semantic search
Hugging Facemaidalun1020/bce-embedding-base_v1768512 tokensChinese, EnglishSentence embeddings, semantic similarity
OpenAItext-embedding-ada-00215368192 tokensMultilingualGeneral text embeddings, semantic search
OpenAItext-embedding-3-small10248192 tokensMultilingualGeneral text embeddings, semantic search
OpenAItext-embedding-3-large20488192 tokensMultilingualHigh-quality embeddings, cross-lingual tasks
Cohereembed-multilingual-v3.01024512 tokensMultilingualSemantic search, retrieval-augmented generation
Cohereembed-english-light-v3.0384512 tokensEnglishSemantic search, retrieval-augmented generation