少点错误 前天 21:32
I am worried about near-term non-LLM AI developments
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章介绍了AI持续学习的研究进展,指出当前主流的离线、非序列训练模式不适合通用人工智能(AGI)的发展,而持续在线、序列学习的模型更具潜力。文章重点介绍了Hierarchical Reasoning Model、Energy-based Model等新型架构,并预测未来6个月将出现性能媲美GPT-3但无需预训练的小型自然语言模型。文章还探讨了持续学习模型的安全性和对齐问题,建议将部分资源转向该领域的研究。

🔍 文章指出当前主流的AI训练模式采用离线、非序列的方式,这与人类在线、序列的学习方式存在显著差异,不利于发展通用人工智能(AGI)。

🧠 持续在线、序列学习的模型更具潜力,因为它们能够像人类一样不断从新数据中学习并更新权重,避免了灾难性遗忘问题。

📈 文章重点介绍了Hierarchical Reasoning Model、Energy-based Model等新型架构,这些模型在ARC-AGI、长上下文视频生成等任务上取得了显著成果。

🚀 文章预测未来6个月将出现性能媲美GPT-3但无需预训练的小型自然语言模型,并认为该模型可能来自专注于持续学习研究的较小实验室。

🤔 文章探讨了持续学习模型的安全性和对齐问题,建议将部分资源转向该领域的研究,并探讨通过交互式训练对模型进行对齐的可能性。

Published on July 31, 2025 1:15 PM GMT

TL;DR

I believe that:

Overview

Almost all frontier models today share two major features in their training regime: they are trained offline and out of sequence

These features of training regimes make sense if you believe the classic function approximation or statistical approximation explanation of machine learning. In this story the model is meant to learn some fixed "target distribution" or "target function" by sampling i.i.d. data points from the training set. The model is then tested on a holdout "test set" which contains new input-output pairs from the same target distribution or function. If the model generalises across the train set and test set, it is considered a good model of the target.

For many reasons, this story makes no sense when applied to the idea of AGI or trying to develop an ML model that is good at navigating the real world. Humans do not spend the first few years of their lives in a sensory deprivation tank, getting random webcam footage from different places on earth before they stop learning forever and are "deployed" into reality. Furthermore, if your plan is to learn the fixed distribution of all possible english sentences, you will naturally need a representative sample of... all possible english sentences. This is not how humans acquire language skills either, and explains why current ML approaches to natural language generation are becoming prohibitively expensive.

Most of us would agree that humans learn continuously, meaning that they learn online and in sequence. Instead of seeing a wide "context" made up of randomly sampled data from all across the internet, we have a very narrow "context" focused on the here and now. To make up for this, we are able to leverage our memories of the immediate and distant past to predict the future. In effect we live in one continuous learning "episode" that lasts from the moment we are born to the moment we die. Naturally, AI researchers have tried to find ways to replicate this in ML models.

The Agenda I am Worried About

I think that the AI safety community has seriously overindexed on LLMs and ChatGPT-style model safety. This is a reasonable choice, because LLMs are a novel, truly impressive, and promising line of AI development. However, in my opinion research into online in-sequence learning has the potential to lead to human-level AGI much more quickly. A single human brain has the energy demands of a lightbulb, instead of the energy demands for all the humans in Wyoming.

I am not alone in this belief. Research into online in-sequence learning has focused around small model, RNN-like approaches which do not use backpropagation through time. Instead, models must update their weights online and generalise based on only one input at a time, forcing them to learn how to leverage their memory/hidden state to predict future data points if they wish to be effective. By contrast, transformers are explicitly encouraged to memorise surface-level patterns to blindly apply to large blocks of context, instead of internalising the content and using it to predict future inputs.

Some notable papers applying this research include the Hierarchical Reasoning Model, Energy-based Models, ARC-AGI without pretraining, and Test Time Training. Some of these techniques (like Test Time Training or Energy-based Models) augment existing transformer architectures, while others represent entirely novel architectures like the ARC-AGI with no pretraining model and the Hierarchical Reasoning Model. These models share the same idea of getting more information out of each data point than a single backpropagation pass can extract. For example, Test Time Training uses a neural network as its hidden state. It also has an internal update step where key information contained in any incoming data point is compressed into the weights of the hidden state network. ARC-AGI without pretraining trains a new network on each data point (a single puzzle in the ARC-AGI corpus), again aiming to get some compressed representation of the key structural information contained in that puzzle. The Hierarchical Reasoning Model and Energy-based Model iterate on their internal representations either for some fixed number of cycles or until some convergence threshold is reached. That way they can extract maximum information from each data point and give themselves more "thinking time" compared to transformers, which must output the next token immediately after one forward pass. The Hierarchical Reasoning Model also uses higher and lower level recurrent modules to separate cognition into low level execution steps and high level planning steps.

So far, research within this track has produced strong/claimed-to-be-SOTA results for ARC-AGI 1, ARC-AGI 2 (along with Sudoku and Maze-solving), and long-context video generation. These models excel at tasks current frontier LLMs or reasoning models struggle at, or radically improve the otherwise lacklustre performance of standard LLMs. Despite differences in implementation, models in this line of research learn online, take less data to train, have smaller parameter counts, and have better bootstrapping performance (generalising based on a limited number of data points). Many of them also claim to be inspired by brain-like architectures or how humans learn. I think that of the approaches I have listed above the Hierarchical Reasoning Model is the most promising candidate to come out of this line of research so far.

Concrete Predictions

I believe that within 6 months this line of research will produce a small natural-language capable model that will perform at the level of a model like GPT-3, but with improved persistence and effectively no "context limit" since it is constantly learning and updating weights. It is likely that this model will not come from an existing major frontier lab, but rather a smaller lab focused on this line of research like Sapient (who developed the Hierarchical Reasoning Model). The simplest case would be something like "We have adapted the HRM for natural language tasks and scaled it up, and it just works".

I believe that further development of this research will produce models that fulfill most of the criteria we associate with "AGI". In general I define this as a model that learns continuously and online from new data, generalises efficiently to new domains while avoiding catastrophic forgetting, and is skilled in a wide variety of tasks associated with human intelligence: natural language generation and understanding, pattern matching, problem solving, planning, playing games, scientific research, narrative writing etc.

Concretely, these are the developments I am predicting within the next six months (i.e. before Feb 1st 2026):

Bonus points:

What I think we should do



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI持续学习 Hierarchical Reasoning Model Energy-based Model AGI 在线学习
相关文章