MarkTechPost@AI 2024年12月26日
Tencent Research Introduces DRT-o1: Two Variants DRT-o1-7B and DRT-o1-14B with Breakthrough in Neural Machine Translation for Literary Texts
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

腾讯研究推出了DRT-o1系统,包括DRT-o1-7B和DRT-o1-14B两个变体,专注于解决神经机器翻译在文学文本中处理隐喻和明喻等复杂语言的挑战。该系统基于Qwen2.5架构,采用多智能体框架,通过翻译器、顾问和评估器之间的迭代协作,提升翻译质量。DRT-o1在文学翻译方面取得了显著进展,超越了现有模型,尤其是在处理文化和上下文细微差别方面。实验结果表明,DRT-o1在BLEU和CometScore指标上均有显著提高,证明了其在复杂语言翻译中的有效性和效率。

📚 DRT-o1系统专注于文学领域,利用Project Gutenberg中的400本公共领域书籍,提取并筛选出包含隐喻和明喻的63,000个句子,用于训练。

🤖 该系统采用多智能体框架,包含翻译器、顾问和评估器三个角色,通过迭代优化翻译结果,确保输出质量。

📈 DRT-o1-7B的BLEU得分提高了8.26分,CometScore提高了3.36分;DRT-o1-14B的BLEU得分提高了7.33分,CometScore提高了1.66分,显著优于其对应的Qwen2.5模型。

✨ 该系统集成了GPT-4o,确保了翻译的流畅性和可读性,进一步提高了机器翻译的质量。

🏆 DRT-o1-7B甚至超越了更大的QwQ-32B模型,在BLEU得分上高出7.82分,CometScore高出1.46分,展示了其在处理复杂文学内容方面的可扩展性和效率。

Neural machine translation (NMT) is a sophisticated branch of natural language processing that automates text conversion between languages using machine learning models. Over the years, it has become an indispensable tool for global communication, with applications spanning diverse areas such as technical document translation and digital content localization. Despite its advancements in translating straightforward text, NMT faces persistent challenges in handling literary content rich in metaphors and similes. These expressions carry deep cultural and contextual nuances, making their translation far more complex. Conventional systems often resort to literal translations, which can fail to preserve the intended meaning and cultural essence, particularly in literature, where semantics are intertwined with artistic and emotional undertones.

Translating idiomatic expressions and metaphorical content involves unique difficulties stemming from their reliance on cultural interpretation. Literal translations of such constructs often lead to a loss of nuance, rendering the output confusing or meaningless to native speakers. This issue persists even with the most advanced NMT systems, designed to excel in tasks involving structured or technical text but falter when interpreting abstract and figurative language. Human translators invest considerable effort in reinterpreting these expressions to ensure they align with the target audience’s cultural framework while retaining the original intent. Bridging this gap in automated systems requires a novel approach capable of mimicking this human adaptability.

Existing NMT tools leverage supervised fine-tuning (SFT) techniques to enhance translation capabilities. These tools typically rely on datasets optimized for technical or straightforward text, such as manuals or academic papers. However, their performance diminishes when dealing with metaphorical or idiomatic language. Systems like Qwen2.5 and Marco-O1 improve accuracy and fluency for basic translations but remain ill-equipped to handle the layered complexities of literary language. For instance, Qwen2.5-7B achieves a BLEU score of 27.02, and Qwen2.5-14B improves this to 30.23, yet neither comes close to meeting the high expectations of literary translation where context and nuance are paramount.

Researchers from Tencent Inc. have developed an innovative system called DRT-o1 to overcome these limitations. It comprises of two variants:

    DRT-o1-7B DRT-o1-14B

They are built upon the Qwen2.5 backbones and integrate a novel multi-agent framework to address the intricacies of metaphorical and idiomatic translation. The researchers focused on literature as their primary domain, mining approximately 400 public-domain English books from Project Gutenberg. They extracted 577,600 sentences and filtered them to retain only 63,000 containing similes and metaphors. These sentences were deemed suitable for what the researchers describe as “long thought” processes in machine translation. Unlike previous approaches, the DRT-o1 system relies on a collaborative method involving three agents: 

    A translatorAn advisorAn evaluator 

Each agent iteratively refines the translation, ensuring that every output improves upon the last.

The multi-agent framework at the core of DRT-o1 begins with identifying key terms in a source sentence. These terms are translated individually to ensure contextual accuracy. The framework then generates a preliminary translation, which undergoes multiple refinement loops. During each iteration, the advisor provides feedback on the current translation, and the evaluator assigns a score based on predefined quality metrics. This iterative process continues until the evaluator’s score meets a predefined threshold or the maximum number of iterations is reached. The outputs are then polished for fluency and readability using GPT-4o, creating a final dataset of 22,264 long-thought machine translation samples.

The DRT-o1 system and its variants significantly improve performance over existing NMT models. Experimental results reveal that DRT-o1-7B achieves an 8.26-point increase in BLEU score and a 3.36-point rise in CometScore compared to its Qwen2.5-7B-Instruct counterpart. Similarly, DRT-o1-14B records a BLEU improvement of 7.33 and a CometScore increase of 1.66 over Qwen2.5-14B-Instruct. These results underscore the effectiveness of the multi-agent framework in capturing the subtleties of literary translation. Notably, DRT-o1-7B even outperforms larger models such as QwQ-32B, demonstrating the scalability and efficiency of this system. For example, the 7B variant surpasses QwQ-32B by 7.82 BLEU points and 1.46 CometScore, further establishing its capabilities in handling complex linguistic constructs.

Key takeaways from the research on the DRT-o1:

    The dataset creation involved mining 577,600 sentences from 400 public-domain books, filtering them to 63,000 suitable for long-thought processes.The multi-agent framework employs three roles – translator, advisor, and evaluator – to iteratively refine translations and ensure superior output quality.DRT-o1-7B improved its BLEU by 8.26 points, while DRT-o1-14B recorded a 7.33-point increase, showcasing the system’s ability to outperform existing models.The integration of GPT-4o ensures fluency and readability, further enhancing the quality of machine translations.DRT-o1-7B outperformed the larger QwQ-32B model by 7.82 BLEU points, highlighting its scalability and efficiency in translating complex literary content.

In conclusion, the DRT-o1 system and its variants (DRT-o1-7B and DRT-o1-14B) represent a transformative approach to neural machine translation. The researchers have addressed long-standing challenges by focusing on literary language and integrating a sophisticated multi-agent framework. The iterative refinement process preserves the meaning and cultural context of metaphors and similes and achieves performance metrics that surpass state-of-the-art models. This work underscores the potential of long-chain reasoning in enhancing NMT, providing a scalable and effective solution for translating nuanced text with precision and cultural sensitivity.


Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

Trending: LG AI Research Releases EXAONE 3.5: Three Open-Source Bilingual Frontier AI-level Models Delivering Unmatched Instruction Following and Long Context Understanding for Global Leadership in Generative AI Excellence….

The post Tencent Research Introduces DRT-o1: Two Variants DRT-o1-7B and DRT-o1-14B with Breakthrough in Neural Machine Translation for Literary Texts appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

神经机器翻译 多智能体框架 文学翻译 DRT-o1 深度学习
相关文章