MarkTechPost@AI 06月06日 12:35
Alibaba Qwen Team Releases Qwen3-Embedding and Qwen3-Reranker Series – Redefining Multilingual Embedding and Ranking Standards
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

阿里巴巴的Qwen团队推出了Qwen3-Embedding和Qwen3-Reranker系列模型,为多语言文本嵌入和相关性排序设立了新标准。这些模型基于Qwen3基础模型构建,提供0.6B、4B和8B三种参数规模版本,支持119种语言,是目前最具通用性和高性能的开源解决方案之一。它们在语义检索、分类、RAG、情感分析和代码搜索等应用场景中表现出色,并已在Hugging Face、GitHub和ModelScope上开源,同时通过阿里云API提供服务。

🌍 Qwen3-Embedding系列模型基于Transformer架构,采用因果注意力机制,通过提取[EOS]标记对应的隐藏状态来生成嵌入。它们具备指令感知能力,通过将输入查询格式化为“{instruction} {query}<|endoftext|>”来实现任务条件下的嵌入。

🛠️ Reranker模型采用二元分类格式进行训练,使用基于token似然的评分函数,以指令引导的方式判断文档-查询的相关性。

⚙️ Qwen3模型的训练流程包括大规模弱监督、监督微调和模型融合三个阶段。大规模弱监督阶段生成1.5亿个合成训练对,覆盖多种语言和任务;监督微调阶段使用1200万高质量数据对;模型融合阶段通过球形线性插值(SLERP)确保模型的鲁棒性和泛化能力。

🏆 在多项基准测试中,Qwen3-Embedding和Qwen3-Reranker系列展现出强大性能。Qwen3-Embedding-8B在MMTEB上平均任务得分为70.58,超过Gemini和GTE-Qwen2系列;在MTEB-Code上达到80.68,在代码检索和Stack Overflow问答等应用中表现出色。

🚀 Qwen3-Reranker-0.6B已超越Jina和BGE rerankers,而Qwen3-Reranker-8B在MTEB-Code上达到81.22,在MMTEB-R上达到72.94,表现出最先进的水平。

Text embedding and reranking are foundational to modern information retrieval systems, powering applications such as semantic search, recommendation systems, and retrieval-augmented generation (RAG). However, current approaches often face key challenges—particularly in achieving both high multilingual fidelity and task adaptability without relying on proprietary APIs. Existing models frequently fall short in scenarios requiring nuanced semantic understanding across multiple languages or domain-specific tasks like code retrieval and instruction following. Moreover, most open-source models either lack scale or flexibility, while commercial APIs remain costly and closed.

Qwen3-Embedding and Qwen3-Reranker: A New Standard for Open-Source Embedding

Alibaba’s Qwen Team has unveiled the Qwen3-Embedding and Qwen3-Reranker Series—models that set a new benchmark in multilingual text embedding and relevance ranking. Built on the Qwen3 foundation models, the series includes variants in 0.6B, 4B, and 8B parameter sizes and supports a wide range of languages (119 in total), making it one of the most versatile and performant open-source offerings to date. These models are now open-sourced under the Apache 2.0 license on Hugging Face, GitHub, and ModelScope, and are also accessible via Alibaba Cloud APIs.

These models are optimized for use cases such as semantic retrieval, classification, RAG, sentiment analysis, and code search—providing a strong alternative to existing solutions like Gemini Embedding and OpenAI’s embedding APIs.

Technical Architecture

Qwen3-Embedding models adopt a dense transformer-based architecture with causal attention, producing embeddings by extracting the hidden state corresponding to the [EOS] token. Instruction-awareness is a key feature: input queries are formatted as {instruction} {query}<|endoftext|>, enabling task-conditioned embeddings. The reranker models are trained with a binary classification format, judging document-query relevance in an instruction-guided manner using a token likelihood-based scoring function.

The models are trained using a robust multi-stage training pipeline:

    Large-scale weak supervision: 150M synthetic training pairs generated using Qwen3-32B, covering retrieval, classification, STS, and bitext mining across languages and tasks.Supervised fine-tuning: 12M high-quality data pairs are selected using cosine similarity (>0.7), fine-tuning performance in downstream applications.Model merging: Spherical linear interpolation (SLERP) of multiple fine-tuned checkpoints ensures robustness and generalization.

This synthetic data generation pipeline enables control over data quality, language diversity, task difficulty, and more—resulting in a high degree of coverage and relevance in low-resource settings.

Performance Benchmarks and Insights

The Qwen3-Embedding and Qwen3-Reranker series demonstrate strong empirical performance across several multilingual benchmarks.

For reranking:

Ablation studies confirm the necessity of each training stage. Removing synthetic pretraining or model merging led to significant performance drops (up to 6 points on MMTEB), emphasizing their contributions.

Conclusion

Alibaba’s Qwen3-Embedding and Qwen3-Reranker Series present a robust, open, and scalable solution to multilingual and instruction-aware semantic representation. With strong empirical results across MTEB, MMTEB, and MTEB-Code, these models bridge the gap between proprietary APIs and open-source accessibility. Their thoughtful training design—leveraging high-quality synthetic data, instruction-tuning, and model merging—positions them as ideal candidates for enterprise applications in search, retrieval, and RAG pipelines. By open-sourcing these models, the Qwen team not only pushes the boundaries of language understanding but also empowers the broader community to innovate on top of a solid foundation.


Check out the Paper, Technical details, Qwen3-Embedding and Qwen3-Reranker. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 95k+ ML SubReddit and Subscribe to our Newsletter.

The post Alibaba Qwen Team Releases Qwen3-Embedding and Qwen3-Reranker Series – Redefining Multilingual Embedding and Ranking Standards appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Qwen3 文本嵌入 多语言 开源
相关文章