少点错误 07月23日 05:17
Translating Everything with LLMs
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了使用大型语言模型(LLMs)将英文AI安全领域的重要论文、文章和讨论翻译成多种主要语言的可行性与价值。作者通过Claude Sonnet 4进行了一次大规模翻译实验,将其翻译成包括中文、韩语、法语、德语等在内的多种语言,并对翻译质量、术语处理(如AGI、alignment)以及潜在的挑战(如token限制、翻译的自然度)进行了评估。实验发现,虽然LLMs在传达核心信息方面表现不错,但要达到母语者般自然流畅的写作风格仍有距离。该方法有望以较低成本加速AI安全知识的全球化传播,但仍需人工校对和术语标准化来提升翻译质量。

💡 LLMs翻译AI安全内容可降低成本并扩大全球影响力:作者认为,利用LLMs进行多语言翻译,可以显著降低AI安全知识的传播成本,使更多非英语国家的人们能够接触和理解AI安全领域的前沿研究和讨论,促进全球对AI风险的共同认知和应对。

🔄 术语翻译是关键挑战:在翻译过程中,“AGI”等缩略语以及“alignment”等核心概念的翻译存在困难。LLMs在处理这些术语时,有时会选择直译或使用通用词汇,而非最贴切或最被广泛接受的专业术语,这可能影响理解的准确性,例如“alignment”在不同语言中的翻译差异很大。

✍️ 翻译质量与自然度仍待提升:尽管LLMs能够生成基本准确的翻译,但翻译出的文本在自然流畅度和可读性方面仍有不足,有时会显得生硬或“没有灵魂”。作者通过调整提示词和使用词汇表等方法尝试改进,但要达到母语般的写作水平仍是长远目标。

💰 Token限制影响翻译完整性:在翻译过程中,不同语言的token数量存在差异,导致部分翻译因超出API设置的token限制而被截断。这不仅影响了翻译的完整性,也可能暗示了不同语言在模型处理效率上的不均衡。

📈 结合人工校对与专业词汇表可优化翻译:为了克服LLMs翻译的局限性,作者建议结合人工校对和预先整理的专业词汇表,以提高翻译的准确性和专业性。一个精心维护的词汇表可以跨越多种语境,显著提升翻译的质量。

Published on July 22, 2025 9:13 PM GMT

There are a lot of AI Safety papers, essays, think pieces, discussions, all written in English. What if we used LLMs to translate them into every major language?

Here I lay out briefly why I think this could be a good idea, and then I attempt to have Claude do some mass translating for me, for better or worse. 

Why do this

Quick look at the cost/benefit:

Cost

If an LLM can do this well, possibly this costs very little? A one person startup could presumably provide translation as a service, and cheaply make things like AI 2027, which are intended for mass consumption, natively multilingual from the start.

Benefit

It seems likely that other countries might also have an influence on the future! The European Union is passing major legislation on AI, the semiconductor supply chain is spread across the globe (with multiple non-Anglosphere choke points), and the UN can presumably at least kick up a fuss about things. It also seems plausible that as the AI race heats up, people in countries other than the US (or China) may be especially alarmed and upset by what’s going on, facing serious risks/harms that they had no part in creating. 

So, how hard could it be?

A trial in translation

I had Claude Sonnet 4[1] translate the essay “Keep the Future Human” (decently long, lots of jargon) into the following languages: Mandarin(simplified[2]), Korean, Dutch, Japanese, Hebrew, Malay, French, Russian, German, Italian, Polish, Turkish, Spanish, Danish, Greek, Arabic, and Latvian.[3] I’m fluent in Dutch, and can kind of read Spanish, but that’s about it for me personally, so it felt like a big leap of faith. I did mess around a little bit with the COT prompts to improve the Dutch translation, with the hope that those improvements would generalize to the other languages as well. 

Claude Code wrote me a basic pipeline where Claude Sonnet 4 would:

    Read the full essay[4] and write a short summary and glossary of terms.Translate the essay one chapter at a time, each time having access to the same summary and glossary.[5] 

These translated chapters were then reassembled into one big document. 

The A-G-I Framework

A core part of the essay is the A-G-I framework, which defines AGI as:

high Autonomy (independence of action), high Generality (broad scope and adaptability), and high Intelligence (competence at cognitive tasks)

This particular acronym, of course, does not often translate well into other languages. For example:

Dutch  ❌ hoge Autonomie (onafhankelijkheid van handelen), hoge Algemene toepasbaarheid (brede reikwijdte en aanpassingsvermogen), en hoge Intelligentie (competentie bij cognitieve taken)

French ✅ haute Autonomie (indépendance d'action), haute Généralité (portée large et adaptabilité), et haute Intelligence (compétence aux tâches cognitives)

Polish ❌ wysokiej Autonomii (niezależności działania), wysokiej Ogólności (szerokiego zakresu i adaptowalności) oraz wysokiej Inteligencji (kompetencji w zadaniach poznawczych

Mandarin ⁉️ 高度自主性(行动独立性)、高度通用性(广泛范围和适应性),以及高度智能(认知任务能力)

 

Generally, Claude chose to keep the acronym AGI without translating it, arguing that in most languages this has become a commonly understood term already. However, in the case where the author breaks down the acronym and redefines it from Artificial General Intelligence to Autonomous General Intelligence, there are actually some tricky choices to make about how to go about adapting what is essentially English wordplay. I’m not really sure how to robustly fix this kind of thing.

I also tried translating the corresponding diagram into Dutch, which had mixed success:

Claude Sonnet 4 decided to make an SVG, which led to more accurate text, but the placement was pretty wonky. GPT-4o generated an image, which looks really great at first glance, but has a few spelling and grammar mistakes. Both of them chose to force “Generality” to “Generaliteit”, which is pretty unnatural, but it does keep the G! I decided to ditch translating the diagrams, and left in the original English version. 

Translating “alignment”

Claude was instructed to make a glossary of important terms and jargon, with their corresponding translations. Looking specifically at the word “alignment”, it usually matched the word used in that language’s Wikipedia entry for AI Alignment (if there was one). An exception was Polish which translated “alignment” to “wyrównanie”, which GPT-4o translated as:

Wyrównanie refers to the act of making things equal, level, or aligned. In physical contexts, it can mean leveling a surface (e.g., terrain or floor). In math or accounting, it can refer to balancing or settling (e.g., wyrównanie rachunków = settling accounts). In social or political discourse, it can imply striving for equality or fairness—e.g., wyrównanie szans = equal opportunity—often with a positive connotation of justice or fairness. It may also be used metaphorically to suggest bringing things into harmony or balance, such as workload, rights, or outcomes. The word does not inherently imply forceful or abrupt correction, but rather a systematic or intended adjustment toward balance.

That seems wrong.[6] For Russian, Claude also went with a similar cognate “выравнивание” (likewise more about leveling/equalizing) despite the Wikipedia entry using “cогласование”. This word (as well as the Latvian cognate, “saskaņošana”, which Claude chose) appears to mean something closer to “coordination” than “alignment”, but it’s hard for me to tell.

Dutch, Hebrew, Malay[7], Italian, Turkish, Danish, Greek, and Latvian did not have Wikipedia entries. For Dutch and German, Claude just used the English term “alignment”, and for Danish it seemed to switch back and forth between “alignment” and a Danish term. All of the other languages seem to have mostly gone with fairly reasonable translations (as far as I can tell!). 

A human curated glossary of terms might make this sort of thing a lot better. While requiring some initial investment of time and effort, the same glossary could be used widely across contexts. 

Some qualitative comments

I asked friends to look over their respective translations, and they commented:

Polish:

It's overall pretty good. It translated the parts about gates weirdly: "closing the gated before AGI", and later "closing the gates for AGI". It should be "to AGI" but it didn't understand the context I think. Also translating AGI into A-O-I was weird - it should somehow relate it to the English acronym.

But apart from occasional awkward things like this, it's quite readable.

French:

It is OK

Content is here, but it is not really pleasant to read. It is a bit "souless" 

In a blind test, I would pbly be able to see it has been AI translated

Greek:

Ok so the literal translation is accurate, but it just doesn’t sound like how a greek person would write

Like it translates idioms or ways of expression word for word, without accounting for the fact that a Greek person wouldn’t talk like that

Looking over the Dutch translation, my assessment is similar. It’s fine, but clunky and not that nice to read, making some weird choices here and there. Claude seems to have erred on the side of being overly literal, sacrificing natural flow to stick closely to the English wording. I tried to fix that with better prompting, but I didn’t try very hard, and I don’t see why this couldn’t be improved a lot. 

I hope the other languages worked out, maybe some of them are radically worse (let me know!). 

Token inequality

Several languages had parts of the translations cut off. Upon closer inspection I noticed that some languages require a lot more tokens, and this led the API to hit the limit I had set of about 16k output tokens per chapter.[8] This is easily fixable, though I wonder what implications the differing amount of tokens has on translation quality.[9] 

Below is each language, with the total number of tokens in Chapter 1:

Cost

I spent roughly 45 dollars in compute credits (Claude Sonnet 4 is a bit pricey). This includes the cost of vibecoding. 

Conclusion

Hard to say how that went! I had GPT-4o critique each of the translations, and it always had something to constructive to say, but assuaged me (perhaps sycophantically) that they were mostly fine. My overall feeling is that passably accurate translation of large volumes of text is probably achievable using better COT, prompting, and human curated glossaries (at least for some languages). 

However, translating text so that the translation is itself good writing is a whole other problem, and that seems much farther off. This seems like the sort of thing LLMs should be good at, and perhaps with much better prompting you could get writing which is much more at home in the target language, and is genuinely enjoyable to read. 

I could imagine translating some other writing that doesn’t necessarily need to hit as high a “style” bar:

I’ve put the results of my experiment up on Github, as well as all the code that I used Claude used. If you can read one of the languages I tried, please let me know how it went! 

P.S. I let Claude write my last commit message, and Claude added itself as a contributor!

  1. ^

    I realize this might be a bit overpowered, but since I was jumping blind into languages I didn’t know, I figured I’d better play it safe.

  2. ^

    China uses simplified characters, while Taiwan uses traditional Chinese characters. Apparently translating from traditional to simplified can be done with a simple mapping table, but the other direction requires the context to resolve ambiguities.

  3. ^

    Chosen very roughly for either having a significant role in the semiconductor supply chain according to a GPT-4o generated summary of this report, being on the 2026 UN Security Council, or being in the top 10 NATO countries by military spending.

  4. ^

    I decided to use Obsidian Web Clipper to scrape the website, and do everything in Markdown for simplicity.

  5. ^

    Maybe I should have just tried it without chucking (I think the context window was large enough). On the other hand, this could be a useful datapoint for much longer pieces of text that have to be chunked. 

  6. ^

    My Polish speaking friend explained that “alignment” is notoriously hard to translate into Polish. Polish EA Slack had a big debate about it, settling on “dostosowywanie” which translates roughly to “the process of making something suitable or appropriate for a specific purpose, condition, or need”. This was Claude’s second choice, including it in the glossary, but it never actually used it in the final translation. In a government document about alignment, they also used “wychowywanie”, or “the long-term process of shaping a person’s behavior, character, values, and social functioning—primarily through upbringing and education”, which is also an interesting choice. The Wikipedia entry used “zgodność” which translates to “compliance or consistency”. 

  7. ^

    While Malay did not have a Wikipedia entry, Indonesian did, which seems to have translated the title “AI Alignment” to “AI Control Problem”, which is the term used more commonly before “alignment” became popular (The Russian and Arabic articles did something similar). In the article, however, they use the term “penyelarasan” for “alignment” rather than what Claude selected, which was “penjajaran”. It seems that Claude’s choice might be overly literal, referring more to physical/spatial alignment, but I’m not sure, and though Malay and Indonesian are very similar it could just be a difference in connotation between the languages. 

  8. ^

    This was roughly the highest I could go before Anthropic wanted me to stream the data, which I couldn’t be bothered to do.

  9. ^

    Probably token efficiency correlates strongly with how much of that language was in the pretraining data, though the data used for training the tokenizer is often different.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI安全 大型语言模型 多语言翻译 术语翻译 AI伦理
相关文章