MarkTechPost@AI 07月25日 12:00
Unsupervised System 2 Thinking: The Next Leap in Machine Learning with Energy-Based Transformers
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

人工智能研究正从模式识别迈向具备类人复杂推理能力。最新的Energy-Based Transformers(EBTs)通过其独特的架构和训练方式,旨在实现“System 2 Thinking”(第二系统思维),即机器的深度、分析性思考过程,且无需特定领域监督。与依赖快速直觉的System 1不同,EBTs通过能量最小化优化过程进行推理,能够动态分配计算资源,自然地建模不确定性,并进行自我验证。这种方法使其在语言和视觉任务上表现出色,且比现有Transformer模型更具扩展性和泛化能力,为构建更强大、灵活的AI系统铺平了道路。

🧠 **EBTs实现“System 2 Thinking”:** Energy-Based Transformers(EBTs)通过引入能量函数和基于能量最小化的迭代优化过程,使机器能够进行类似人类的深度、分析性推理,而非仅仅依赖模式匹配。这标志着AI从“System 1 Thinking”(快速、直觉)向“System 2 Thinking”(缓慢、分析)的转变。

💡 **无需监督的泛化能力:** EBTs的核心优势在于其无监督学习能力,这意味着它们可以在没有特定领域知识或人工奖励信号的情况下,自主学习复杂的推理过程。这种方法使其能够处理更广泛、更具挑战性的任务,并能更好地泛化到未见过的数据或情境中。

⚙️ **动态计算与不确定性建模:** EBTs能够根据任务的复杂性和预测的不确定性,动态地调整计算资源的分配,即“思考”的深度。同时,它们能自然地建模和量化自身的不确定性,这对于处理连续域(如图像)或需要高可靠性判断的任务至关重要。

🚀 **优于传统Transformer的性能:** 实验表明,EBTs在语言和视觉任务上,通过增加“思考”时间,能够显著提升下游任务的性能。更重要的是,它们在数据、计算和模型规模方面的训练效率更高,且在任务难度增加或超出分布时,泛化能力表现出更强的提升,显示出其优越的可扩展性和鲁棒性。

Artificial intelligence research is rapidly evolving beyond pattern recognition and toward systems capable of complex, human-like reasoning. The latest breakthrough in this pursuit comes from the introduction of Energy-Based Transformers (EBTs)—a family of neural architectures specifically designed to enable “System 2 Thinking” in machines without relying on domain-specific supervision or restrictive training signals.

From Pattern Matching to Deliberate Reasoning

Human cognition is often described in terms of two systems: System 1 (fast, intuitive, automatic) and System 2 (slow, analytical, effortful). While today’s mainstream AI models excel at System 1 thinking—rapidly making predictions based on experience—most fall short on the deliberate, multi-step reasoning required for challenging or out-of-distribution tasks. Current efforts, such as reinforcement learning with verifiable rewards, are largely confined to domains where correctness is easy to check, like math or code, and struggle to generalize beyond them.

Energy-Based Transformers: A Foundation for Unsupervised System 2 Thinking

The key innovation of EBTs lies in their architectural design and training procedure. Instead of directly producing outputs in a single forward pass, EBTs learn an energy function that assigns a scalar value to each input-prediction pair, representing their compatibility or “unnormalized probability.” Reasoning, in turn, becomes an optimization process: starting from a random initial guess, the model iteratively refines its prediction through energy minimization—akin to how humans explore and check solutions before committing.

This approach allows EBTs to exhibit three critical faculties for advanced reasoning, lacking in most current models:

    Dynamic Allocation of Computation: EBTs can devote more computational effort—more “thinking steps”—to harder problems or uncertain predictions as needed, instead of treating all tasks or tokens equally.Modeling Uncertainty Naturally: By tracking energy levels throughout the thinking process, EBTs can model their confidence (or lack thereof), particularly in complex, continuous domains like vision, where traditional models struggle.Explicit Verification: Each proposed prediction is accompanied by an energy score indicating how well it matches the context, enabling the model to self-verify and prefer answers it “knows” are plausible.

Advantages Over Existing Approaches

Unlike reinforcement learning or externally supervised verification, EBTs do not require hand-crafted rewards or extra supervision; their system 2 capabilities emerge directly from unsupervised learning objectives. Moreover, EBTs are inherently modality-agnostic—they scale across both discrete domains (like text and language) and continuous ones (such as images or video), a feat beyond the reach of most specialized architectures.

Experimental evidence shows that EBTs not only improve downstream performance on language and vision tasks when allowed to “think longer,” but also scale more efficiently during training—in terms of data, compute, and model size—compared to state-of-the-art Transformer baselines. Notably, their ability to generalize improves as the task becomes more challenging or out-of-distribution, echoing findings in cognitive science about human reasoning under uncertainty.

A Platform for Scalable Thinking and Generalization

The Energy-Based Transformer paradigm signals a pathway toward more powerful and flexible AI systems, capable of adapting their reasoning depth to the demands of the problem. As data becomes a bottleneck for further scaling, EBTs’ efficiency and robust generalization can open doors to advances in modeling, planning, and decision-making across a wide array of domains.

While current limitations remain—such as increased computational cost during training and challenges with highly multi-modal data distribution—future research is poised to build on the foundation laid by EBTs. Potential directions include combining EBTs with other neural paradigms, developing more efficient optimization strategies, and extending their application to new multimodal and sequential reasoning tasks.

Summary

Energy-Based Transformers represent a significant step towards machines that can “think” more like humans—not simply reacting reflexively, but pausing to analyze, verify, and adapt their reasoning for open-ended, complex problems across any modality.


Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project.

Meet the AI Dev Newsletter read by 40k+ Devs and Researchers from NVIDIA, OpenAI, DeepMind, Meta, Microsoft, JP Morgan Chase, Amgen, Aflac, Wells Fargo and 100s more [SUBSCRIBE NOW]

The post Unsupervised System 2 Thinking: The Next Leap in Machine Learning with Energy-Based Transformers appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Energy-Based Transformers 人工智能 机器推理 无监督学习 深度学习
相关文章