Compare text and multimodal embedding models from Jina, Qwen, Google, OpenAI and more. Find the best model for semantic search, RAG, and recommendation systems.
Universal embedding model for multimodal and multilingual retrieval. Supports both single-vector and...
Multilingual embeddings with task-specific LoRA adapters. Optimized for text retrieval, semantic mat...
State-of-the-art multilingual text embedding model. #1 on MTEB multilingual leaderboard. Supports 10...
Google's text embedding model for semantic search, classification, and clustering. Supports flexible...
SOTA BERT-large sized embedding model. Outperforms OpenAI text-embedding-3-large and matches models ...
Multimodal embedding model for text, images, and PDFs. Supports flexible dimensions and 128K context...
English-optimized text and image embedding model. Fast and accurate for classification and retrieval...
Multilingual embedding model supporting 100+ languages. Ideal for cross-lingual retrieval and classi...
State-of-the-art general-purpose and multilingual embedding. Outperforms OpenAI-v3-large by 9.74% ac...
Top-ranked embedding model on MTEB English and Chinese. Built on Qwen2-7B with bidirectional attenti...
Top retrieval model on MTEB, ranking #1 with 60.2 score. Built on Mistral-7B with advanced data refi...
Late interaction retriever with excellent multilingual and cross-lingual performance. Fast inference...
1.74B-parameter multimodal omni embedding model accepting text, images, video, and audio with shared...
1.04B-parameter compact multimodal omni embedding model accepting text, images, video, and audio. Sh...
0.95B-parameter text-matching-targeted variant of the v5 omni-nano family. Accepts text, images, vid...
677M-parameter multilingual text embedding model scoring 71.7 on MTEB English v2 and 67.7 on MMTEB. ...
239M-parameter lightweight multilingual text embedding model scoring 71.0 on MTEB English v2 and 65....
677M-parameter text-matching-targeted variant of v5-text-small. Optimized for symmetric pairwise sim...
Embedding models convert text, images, and other data into numerical vectors that capture semantic meaning. These vectors enable semantic search, recommendation systems, and Retrieval-Augmented Generation (RAG) for AI applications.
Higher dimensions capture more nuance but require more storage. Range from 384D to 8192D.
Maximum tokens the model can process in a single input. Critical for long documents.
Support for text, images, PDFs, and more. Multimodal models enable unified search.
Dense (most common), sparse (keyword-focused), late interaction (ColBERT), and more.
Learn more about embedding models and vector search from these authoritative sources: