少点错误 06月05日 03:17
A Technique of Pure Reason
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了大型语言模型(LLMs)未来的发展方向,预测LLMs将不再专注于成为知识渊博的聊天机器人,而是演变为更高效的、侧重推理能力的纯推理模型。这种转变将通过工具使用和其他辅助手段来实现。文章分析了当前LLMs在知识存储方面的低效性,并提出了通过解耦推理能力和记忆能力来提高效率的多种技术途径,包括强化学习、专家混合模型和知识蒸馏。最后,文章讨论了这种转变对AI安全性的影响,强调了纯推理模型在可解释性方面的优势。

🧠 当前LLMs的知识存储效率较低:现有LLMs的参数很大一部分用于存储事实知识,导致训练和推理成本高昂,并可能影响泛化能力。文章认为,模型对知识的记忆并非必要,应该将重点放在推理能力上。

🛠️ 纯推理模型的实现技术:文章提出了几种实现纯推理模型的技术,包括基于强化学习的推理方法(如链式思考)、专家混合模型(MoE)以及知识蒸馏。这些技术旨在提高模型的效率,并解耦推理能力和记忆能力。

💡 知识获取与工具使用的重要性:纯推理模型需要依赖外部知识来源,例如检索增强生成、工具使用和子代理。随着模型在大型上下文窗口中准确关注信息的能力提高,这些技术将变得更加重要。

🛡️ 纯推理模型对AI安全性的影响:纯推理模型在AI控制方面具有优势,因为它们依赖外部知识输入,使得LLMs的思考过程更加透明和可解释。这有助于人类更好地理解和控制AI的行为。

Published on June 4, 2025 7:07 PM GMT

Looking a little ahead into the future, I think LLMs are going to stop being focused on knowledgeable, articulate[1] chatbots, but instead be more efficient models that are weaker in these areas than current models, but relatively stronger at reasoning, a pure-reasoner model. The rest will be bolted on via tool-use and other scaffolding[2].

The Current Idiom is Inefficient

Chatbots make a great replacement for Google searches because they know a lot about everything. But that knowledge comes at a price. The majority of an LLM's parameters are thought to be spent on storing factual knowledge. Parameters are the key determinant of training and inference costs. It likely has a significant cost on data efficiency, and it's possible these extra parameters just hurt generalisation overall.

Memorisation is just not a feature that we need models to have

We keep building bigger models which memorise more and more data largely because we don't know how not to. Every sentence of pre-training simultaneously encourages the model to memorise more factual information, and ekes it towards better high-level thinking and eventually reasoning ability.

But I think that technicality will fall soon, driving towards sleeker agents that fit inside a larger framework[3].

The Technical Path to Pure Reason

The main blocker at the moment is that we don't know how to train models that have good capability without also training them to be great memorisers. There's also a lesser blocker of ensuring that such a model is reliably supplied in context with any information it needs, to compensate for not being able to furnish it itself.

Decoupling Reasoning Capability from Memorisation

While I cannot predict exactly what techniques are likely to work, it's clear there are several productive areas of current research that all lead in the same direction.

Reasoning RL. While I've been using "reasoning" in the colloquial sense, the biggest recent advance in reasoning capabilities comes from reinforcement learning techniques that encourage a reasoning style chain-of-thoughts. This technique is still new enough that there are likely remaining gains[4]

Reasoning models often have very long thinking traces, which makes all those parameters dedicated to memorisation even more expensive.

Mixture of Experts can be viewed as an alternate take on pure reasoners. They cut the parameter count in use at once, bringing similar efficiency gains to separating memorisation. It's unclear at present if MoE can be fairly described as a separation, though.

Distillation allows smaller models to be trained from a larger model. The smaller model often loses some performance - it's not hard to imagine a training regime which preserves reasoning capabilities while losing memorisation capabilities.

Supplying the Pure Reasoning Model with Needed Knowledge

I hardly feel I need to spell this one out. Retrieval Augmented Generation, tool use and subagents are all standards of a modern LLM-based application. These techniques will probably improve over time.

Simultaneously, there is also furious work on architectural improvements to attention. As models get better at attending accurately in large context windows, in-context techniques like the above become relatively stronger compared with parameter-based memorisation.

Once you have pure-reasoning models, it will also make sense for those models to call out to other LLMs that are not as smart, but do have the wide knowledge base. Or possibly these will both be the same model, as a mixture-of-experts where the majority of the model isn't active most of the time.

Implications for Safety

I think this development will be good news, at least for AI control, not AI alignment. Current reasoners show chains of thought that are very useful for inspection. If only we could rely on them being a faithful, complete representation of thought.

If instead we have very smart models that are completely dependent on their scaffolding for providing the facts, then we essentially gain an additional human-readable window into what the LLMs are currently recalling (and thus, thinking about). And unlike chain of thought, it is much harder for pure reasoners to fake - they need this information loaded in order to perform the task.

 

  1. ^

    Most of this essay focuses on knowledge and memorisation rather than language and argumentation skills. It's plausible these and other skills will suffer the same fate.

  2. ^

    For example, Search-R1.

  3. ^

    The bitter lesson suggests that progress will come more from increases in compute than architectural changes like my suggestion. I'm not contradicting that. I just think we've reached the point where models are large enough that there is also value in trimming the fat, so we can spend that compute (and data) on the areas that are most useful.

  4. ^

    I've seen it suggested that RL only elicits existing capabilities. If true, then the ability to improve base models to be better reasoners will be limited to some fixed amount.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

LLM 纯推理 模型效率 AI安全
相关文章