Interconnects 14小时前
Kimi K2 and when "DeepSeek Moments" become normal
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了中国初创公司Moonshot AI发布的Kimi K2模型,该模型在性能上与美国领先模型相当,并采用开放许可。文章指出,Kimi K2的出现预示着西方世界在AI领域可能面临的挑战。Kimi K2在编码和相关任务中表现出色,超越了DeepSeek V3,并成为目前最佳的开放模型。文章还强调了控制前沿模型训练的难度、中国在模型性能上的进步以及西方在开放模型领域的滞后。Kimi K2的发布可能引发西方AI实验室重新思考其研发策略,并推动对开放科学项目的投资。

🚀 Kimi K2是一款开放的混合专家模型(MoE),拥有1万亿参数,其中320亿为活跃参数。它在编码和相关任务中表现出色,性能优于DeepSeek V3。

💡 Kimi K2的训练使用了15.5万亿个tokens,其活跃参数数量与DeepSeek V3/R1相似,后者的训练使用了14.8万亿个tokens。这表明,在没有显著增加计算量的情况下,可以训练出更好的模型,这得益于算法和效率的提升。

🌍 与此同时,西方在开放模型领域的领先地位正在缩小。目前,来自中国的DeepSeek、Moonshot AI和Qwen等公司已经发布了更多有用的模型,且许可更宽松。

⚠️ Kimi K2的发布可能导致西方AI实验室重新思考其研发策略,文章认为,西方需要加大对开放科学项目的资金投入,以保持其在AI研究和开发领域的长期领先地位。

The DeepSeek R1 release earlier this year was more of a prequel than a one-off fluke in the trajectory of AI. Last week, a Chinese startup named Moonshot AI dropped Kimi K2, an open model that is permissively licensed1 and competitive with leading frontier models in the U.S. If you're interested in the geopolitics of AI and the rapid dissemination of the technology, this is going to represent another "DeepSeek moment" where much of the Western world — even those who consider themselves up-to-date with happenings of AI — need to change their expectations for the coming years.

In summary, Kimi K2 shows us that:

Share

Kimi K2, described as an "Open-Source Agentic Model" is a sparse mixture of experts (MoE) model2 with 1T total parameters (~1.5x DeepSeek V3/R1's 671B) and 32B active parameters (similar to DeepSeek V3/R1's 37B). It is a "non-thinking" model with leading performance numbers in coding and related agentic tasks (earning it many comparisons to Claude 3.5 Sonnet), which means it doesn't generate a long reasoning chain before answering, but it was still trained extensively with reinforcement learning. It clearly outperforms DeepSeek V3 on a variety of benchmarks, including SWE-Bench, LiveCodeBench, AIME, or GPQA, and comes with a base model released as well. It is the new best-available open model by a clear margin.

These facts with the points above all have useful parallels for what comes next:

Subscribe now

Kimi K2 is set up for a much slower style "DeepSeek Moment" than the DeepSeek R1 model that came out in January of this year because it lacks two culturally salient factors:

    DeepSeek R1 was revelatory because it was the first model to expose the reasoning trace to the users, causing massive adoption outside of the technical AI community, and

    The broader public is already aware that training leading AI models is actually very low cost once the technical expertise is built up (recall the DeepSeek V3 $5M training cost number), i.e. the final training run is cheap, so there should be a smaller reaction to similar cheap training cost numbers in the Kimi K2 report coming soon.

Still, as more noise is created around the K2 release (Moonshot releases a technical report soon), this could evolve very rapidly. We've already seen quick experiments spin up slotting it into the Claude Code application (because Kimi's API is Claude-compatible) and K2 topping many nice "vibe tests" or creativity benchmarks. There are also tons of fun technical details that I don't have time to go into — from using a relatively unproven optimizer Muon3 and scaling up the self-rewarding LLM-as-a-judge pipeline in post-training. A fun tidbit to show how much this matters relative to the noisy Grok 4 release last week is that Kimi K2 has already surpassed Grok 4 in API usage on the popular OpenRouter platform.

Later in the day on the 11th, following the K2 release, OpenAI CEO Sam Altman shared the following message regarding OpenAI's forthcoming open model (which I previously shared more optimistic thoughts on here) :

we planned to launch our open-weight model next week.

we are delaying it; we need time to run additional safety tests and review high-risk areas. we are not yet sure how long it will take us.

while we trust the community will build great things with this model, once weights are out, they can’t be pulled back. this is new for us and we want to get it right.

sorry to be the bearer of bad news; we are working super hard!

Many attributed this as a reactive move by OpenAI to get out from the shadow of Kimi K2's wonderful release and another DeepSeek media cycle.

Even though someone at OpenAI shared with me that the rumor that Kimi caused the delay for their open model is very likely not true, this is what being on the back foot looks like. When you're on the back foot, narratives like this are impossible to control.

We need leaders at the closed AI laboratories in the U.S. to rethink some of the long-term dynamics they're battling with R&D adoption. We need to mobilize funding for great, open science projects in the U.S. and Europe. Until then, this is what losing looks like if you want The West to be the long-term foundation of AI research and development. Kimi K2 has shown us that one "DeepSeek Moment" wasn't enough for us to make the changes we need, and hopefully we don't need a third.

1

The modified MIT license is somewhat annoying, but technically easy to comply with. These sorts of added terms on marketing make it in conflict with "true open-source principles".

2

Very similar to DeepSeek architecture.

3

Beautiful learning curve.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Kimi K2 Moonshot AI 开放模型 AI竞争 中国AI
相关文章