Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Qwen3-VL-Embedding and Qwen3-VL-Reranker: A Unified Framework for State-of-the-Art Multimodal Retrieval and Ranking

About

In this report, we introduce the Qwen3-VL-Embedding and Qwen3-VL-Reranker model series, the latest extensions of the Qwen family built on the Qwen3-VL foundation model. Together, they provide an end-to-end pipeline for high-precision multimodal search by mapping diverse modalities, including text, images, document images, and video, into a unified representation space. The Qwen3-VL-Embedding model employs a multi-stage training paradigm, progressing from large-scale contrastive pre-training to reranking model distillation, to generate semantically rich high-dimensional vectors. It supports Matryoshka Representation Learning, enabling flexible embedding dimensions, and handles inputs up to 32k tokens. Complementing this, Qwen3-VL-Reranker performs fine-grained relevance estimation for query-document pairs using a cross-encoder architecture with cross-attention mechanisms. Both model series inherit the multilingual capabilities of Qwen3-VL, supporting more than 30 languages, and are released in $\textbf{2B}$ and $\textbf{8B}$ parameter sizes to accommodate diverse deployment requirements. Empirical evaluations demonstrate that the Qwen3-VL-Embedding series achieves state-of-the-art results across diverse multimodal embedding evaluation benchmarks. Specifically, Qwen3-VL-Embedding-8B attains an overall score of $\textbf{77.8}$ on MMEB-V2, ranking first among all models (as of January 8, 2025). This report presents the architecture, training methodology, and practical capabilities of the series, demonstrating their effectiveness on various multimodal retrieval tasks, including image-text retrieval, visual question answering, and video-text matching.

Mingxin Li, Yanzhao Zhang, Dingkun Long, Keqin Chen, Sibo Song, Shuai Bai, Zhibo Yang, Pengjun Xie, An Yang, Dayiheng Liu, Jingren Zhou, Junyang Lin• 2026

Related benchmarks

TaskDatasetResultRank
Object Hallucination EvaluationPOPE--
935
Text-to-Image RetrievalFlickr30K
R@181.9
460
Image-to-Text RetrievalFlickr30K
R@192.9
379
Science Question AnsweringScienceQA
Accuracy92.5
229
Multimodal UnderstandingSEED-Bench
Accuracy78.3
203
Diagram UnderstandingAI2D
Accuracy82.2
167
Text-to-Image RetrievalCOCO
Recall@155.2
130
Image-to-Text RetrievalCOCO
R@169.6
123
Text-to-Image RetrievalFlickr30K-CN
R@178.6
99
Image-to-Text RetrievalFlickr30K-CN
R@191.9
99
Showing 10 of 46 rows

Other info

GitHub

Follow for update