Unite.AI 2024年12月16日
Meta’s COCONUT: The AI Method That Thinks Without Language
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Meta和UC San Diego的研究发现一种新的COCONUT方法,让AI模型以更自然的方式思考。传统方法使AI模型面临诸多问题,而COCONUT给予其自然灵活性,训练过程分阶段,且在多项测试中取得较好成果。

多数语言模型生成的词无推理价值

关键决策点因特定用词需求而受限

模型在语法一致上花费大量算力

COCONUT让AI模型以两种方式思考

COCONUT训练过程分三个阶段

When researchers first discovered that large language models (LLMs) could “think” step by step through chain-of-thought prompting, it was a breakthrough moment – finally, we could peek into the reasoning process of these black boxes. But what if I told you that making AI models think in natural language might be holding them back?

That is what researchers at Meta and UC San Diego have uncovered with their new COCONUT (Chain of Continuous Thought) method.

Imagine trying to solve a complex math problem while being forced to narrate every single step out loud. Annoying, right? Now you're getting close to understanding the core challenge that language models face.

When we make AI models reason through natural language:

The researchers found something interesting in their neuroimaging studies: when humans tackle complex reasoning tasks, the language centers of our brains often remain surprisingly quiet. Yet we have been building AI systems that do the opposite – forcing them to translate every reasoning step into words.

Think about how you solve a puzzle. Your mind probably explores multiple possibilities simultaneously, maintains fuzzy hypotheses, and only crystallizes its thoughts into language when sharing the solution. But traditional chain-of-thought approaches force AI models to verbalize every intermediate step, creating a “linguistic bottleneck.”

This insight led to a compelling question: What if we could let AI models reason in their native “language” – the continuous, high-dimensional space of their hidden states – rather than forcing them to translate everything into tokens?

Understanding COCONUT's Innovation

Picture the difference between speaking your thoughts out loud and the actual mental process happening in your brain. That gap – between verbalized thoughts and neural activity – is exactly what Meta's researchers tapped into with COCONUT.

The real breakthrough of COCONUT lies in how it lets AI models think in two distinct ways, much like how humans do. Think about when you're solving a complex puzzle – you don't narrate every possible move in your head, right? Instead, you:

  1. Absorb the Problem: You take in the information (like reading the puzzle rules)
  2. Think Silently: Your brain explores multiple possibilities without putting them into words
  3. Share the Solution: Only then do you explain your thinking to others

COCONUT gives AI models this same natural flexibility. Instead of forcing them to “speak” every thought out loud (like traditional methods do), it lets them think in their natural neural space – what researchers call the “latent space.”

The model smoothly switches between two modes:

  • When it needs to understand questions or give answers, it uses regular language
  • But for the actual thinking process? It uses pure neural patterns, free from the constraints of words

Image: Meta

The Training Journey

One of the most fascinating aspects of COCONUT is its training curriculum. What makes this one special is how it mirrors natural learning progression. Think about how we teach complex skills – you don't throw someone into the deep end immediately. You build up gradually, adding complexity as they master each level.

The researchers took this exact approach with COCONUT:

Stage 1: The Foundation

First, the model learns like any other AI – through traditional chain-of-thought reasoning. This gives it a solid base understanding.

Stage 2: The Transition

Here is where it gets interesting. Gradually, those written-out reasoning steps get replaced with continuous thoughts. Imagine slowly removing the training wheels, letting the model develop its own internal thinking patterns.

Stage 3: The Balance

Finally, the model learns to seamlessly switch between deep thinking in latent space and communicating its insights in clear language.

During training, the model developed abilities nobody explicitly programmed – like considering multiple reasoning paths simultaneously. This emergent behavior is particularly exciting because it suggests we might be getting closer to more natural forms of AI reasoning. It is these unexpected developments that often lead to the biggest breakthroughs.

Remember those neuroimaging studies I mentioned earlier? They showed that human brains often process complex reasoning tasks without heavily engaging language centers. COCONUT seems to be developing similar patterns – thinking deeply in its native neural space and only converting to language when needed for communication.

The Numbers Tell a Story

A few more key findings stand out from the research:

  • Math Word Problems (GSM8k): Here, COCONUT achieved 34.1% accuracy. While this falls below traditional Chain-of-Thought (42.9%), it's significantly better than baseline approaches.
  • Logical Deduction (ProntoQA): COCONUT hit 99.8% accuracy, edging out traditional Chain-of-Thought's 98.8%. But here's the kicker – it did this while using just 9 tokens compared to CoT's 92.5.
  • Complex Planning (ProsQA): The most impressive results came from this advanced reasoning test. COCONUT achieved 97% accuracy while traditional methods only reached 77.5%. And again, it did this with remarkable efficiency – 14.2 tokens versus 49.4.

What makes these results promising is not just the raw numbers – it is what they reveal about different types of thinking. While COCONUT may still be finding its footing with mathematical reasoning, it excels at tasks requiring complex logical planning and deduction.

COCONUT represents a fundamental rethinking of how AI systems can reason, and it moves us closer to more natural, efficient, and powerful forms of artificial intelligence. The journey from language-based reasoning to continuous thought is a step toward more capable and efficient AI systems.

The post Meta’s COCONUT: The AI Method That Thinks Without Language appeared first on Unite.AI.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

COCONUT AI思考 语言模型 训练过程 成果测试
相关文章