少点错误 03月27日 07:32
Knowledge, Reasoning, and Superintelligence
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了人类和AI的智能,包括结晶智能和流体智能,阐述了其概念、区别及在具体系统中的应用,还提到了知识生产循环以及对未来的设想。

结晶智能与流体智能的概念及区别,如人类中某些活动对两种智能的依赖情况。

AI系统中相关定义,如结晶知识、思考能力等,及二者在AI中的体现。

以AlphaGo和当前LLMs为例,说明知识生产循环在具体系统中的应用。

对未来超级智能系统的设想,如高结晶智能和高流体智能的情况。

Published on March 26, 2025 11:28 PM GMT

What makes people good at solving novel problems? In part, it’s having relevant knowledge. And in part, it’s being able to think flexibly to apply that knowledge. So intelligence isn't a scalar: there are at least two dimensions — often called crystallized intelligence and fluid intelligence.

This is also broadly true of AI. We can see the distinction between language models regurgitating facts and reasoning models tracing out implications. We can see the difference between the performance of AlphaGo’s policynet, and what it achieves with Monte Carlo tree search. And we can imagine this distinction even for superintelligent systems.

This is a conceptual discussion. But the motivation is pragmatic. If we have a better understanding of the high-level workings of intelligence, we can make better predictions about what is required for superhuman performance, or what the dynamics of an intelligence explosion might look like. And this could also help us in thinking about how AI systems should be governed — e.g. should we be giving more attention to questions of who controls the creation and distribution of new knowledge?

Crystallized intelligence vs fluid intelligence

For the purposes of this post, intelligence is the ability to process information in ways that are helpful for achieving goals. Intelligent systems typically have both some crystallized intelligence (a body of implicit or explicit knowledge relevant to the task) and some fluid intelligence (being able to think things through on the fly).

In humans, knowing your multiplication tables, knowing how to ride a bike, or playing bullet chess are all chiefly about crystallized intelligence. But proving novel theorems, designing a bicycle from first principles, or playing slower games of chess all depend more on fluid intelligence.

Of course even in these cases crystallized intelligence is an important ingredient for the ability to think things through — e.g. mathematicians are more likely to be able to prove novel theorems if they have a solid grasp of the existing concepts and theorems, rather than needing to rederive them each time. And since our thinking often depends on our lower-level concepts and heuristics, it’s difficult to give clean definitions separating them.

For AI systems, we can be a little more precise. But first we’ll make a couple of related definitions:

The knowledge production loop

Crystallized intelligence depends crucially on knowledge. But where does the knowledge come from? Of course it may be exogenous to the system (e.g. LLMs functioning as crystallized intelligence based on human-produced data). But something interesting happens in cases where the system can produce its own new knowledge/data — via having new ideas, or via taking actions and observing the results. This gives rise to a loop where improved knowledge leads to stronger crystallized intelligence (and perhaps stronger fluid intelligence), which in turn can produce new higher-quality knowledge:

This is, in essence, a knowledge production engine. It isn’t a closed system — it requires compute to fuel the engine. But with sufficient fuel it may lead to large improvements in performance.

This is perhaps the simplest form of “recursive improvement” — subsequent iterations build on the improvement from earlier steps. And we’ve already seen cases where this kind of engine can produce significantly superhuman systems.

Crystallized intelligence and fluid intelligence in concrete systems

AlphaGo

The creation of the AlphaGo system has steps corresponding to a version of this loop:

    It starts with a database consisting of many expert human gamesThe initial policy network and value network function as a form of crystallized knowledge — they provide useful information giving generalized approximations to what good play looks likeMonte Carlo tree search is a form of thinking algorithm — it can be used to search for good moves (making use of inference compute to make calls to crystallized intelligence to help direct its search)It uses self-play to produce many decision-situations to analyseIt considers and takes actions in these self-play games, and observes the output (i.e. which side wins)This creates many games (eventually with superhuman performance), which are added to the knowledge baseThis improves the policy network and value network in subsequent iterations — and hence the performance of the fluid intelligence

The successor system AlphaZero completes this same loop, but without the initial knowledge base — just the rules of the game (Go or Chess or Shogi) permitting it to make observations. Interestingly, although it builds up a suite of implicit concepts (overlapping with human concepts, but containing new ones), in contrast to the development of the human corpus of expertise, these concepts are purely implicit — they live only in the neural net representing crystallized intelligence, rather than explicit “new ideas” which live in the knowledge base.

Current LLMs

Current LLMs are borderline on completing the knowledge creation loop:

The LLMs of early 2025 are generally strong at crystallized intelligence. They have access to a lot of the data that human civilization has produced, including implicit knowledge about what best practices mean in many many situations.

LLMs are able to apply a certain amount of fluid intelligence, via chain of thought (or scaffolding). At the moment of writing, they’re not always very good at it, but they’re getting better fairly quickly — today’s reasoning models are much stronger than those of a year ago. Some of that strength flows from the kind of knowledge production loop marked:

On the other hand, LLMs aren’t yet at the edge of being able to produce new knowledge which is very helpful by the standards of human civilization. Compared to AlphaGo-like systems, this recursive loop appears to be hamstrung by the lack of clear feedback, and by the breadth of the search space. So it can only improve performance so far — for now.

The future

The distinction between crystallized and fluid intelligence is still meaningful for systems which are superhuman — here are some conceivable (but unrealistic) AI systems which pull out the distinctions:

Crystallized knowledge will remain relevant

It is quite plausible that the basic knowledge production loop could give us transformative AI without any kind of strongly superhuman fluid intelligence. This would involve AI being leveraged to figure out many more things, at scale — the “country of geniuses in a datacentre” — and then the things that it figures out will be the new knowledge worth crystallizing.

Moreover, crystallized knowledge is likely to remain important even for strongly superintelligent systems. Even if you could figure everything out from first principles every time, it’s much less efficient than using cached knowledge [2]— similarly, the first transformative superintelligent systems almost certainly wouldn’t be meaningfully superintelligent if they did have to figure everything out each time.

The fluid intelligence enhancement loop

While the knowledge production loop may drive a large increase in knowledge (and hence crystallized intelligence), classical thinking about an intelligence explosion is more centrally about improvements to fluid intelligence. We might imagine something like this loop:

Even in this case, it seems likely that a lot of the improvements will be driven by the more traditional knowledge production loop — a system might iterate building new knowledge on top of new knowledge for many steps before it reaches an improvement to the algorithms that represents a fundamental upgrade to its fluid intelligence. But the possibility of this further type of upgrade increases the explosive potential of recursive improvement, compared to automated research without this possibility.

So what?

This post has been an exploration of conceptual foundations for thinking about intelligence, and the future of AI. Where can this take us? Ultimately I don’t know; I’m offering up the thinking in part in the hope that others may find useful applications beyond those I’ve considered. But here are some questions that I find my attention drawn towards:

Thanks to Raymond Douglas, Lizka Vaintrob, Tom Davidson, Rudolf Laine, and Andreas Stuhlmüller for helpful comments, and to many more people for helpful conversations.

  1. ^

    Perhaps it would be better to describe things in terms of some ELO-like score, which has a more natural crisp meaning; however that would require a choice of game, and I don’t in this context care enough to try to nail it down.

  2. ^

    This point is not original; Beren makes a similar point about direct and amortized optimization in individual systems, and Rudolf Laine makes an analogous point about amortized optimization at the level of civilization.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

结晶智能 流体智能 知识生产循环 AI系统
相关文章