少点错误 02月19日
Abstract Mathematical Concepts vs. Abstractions Over Real-World Systems
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了纯理论抽象与现实世界抽象之间的区别,例如“向量”与“树”。作者试图形式化这种区分,以便控制超智能体,使其只能思考数学和工程问题,避免对现实世界产生不良影响。文章提出了几种可能的区分方法,包括基于参照完整性、复杂性、因果关系和再现性的定义。其中,基于再现性的定义似乎最有希望,即纯理论概念作为“成分”在许多不同的现实世界抽象中重复出现,而现实世界概念仅出现一次。

💡**纯理论抽象与现实世界抽象的区别:** 纯理论抽象(如向量、博弈论智能体)不涉及特定现实世界系统,适用于不同宇宙;现实世界抽象(如树、人类政府)与地球特定条件相关,属于特定宇宙区域。

🤔**控制超智能体的设想:** 通过移除超智能体思考“未批准”现实世界概念的能力,使其将自身视为非世界的一部分,从而解决定义明确的数学和工程问题,避免其产生现实世界后果。

🔄**基于再现性的定义:** 纯理论抽象作为“成分”在许多不同的现实世界抽象中重复出现(如市场概念),而现实世界抽象(如地球树)仅出现一次,不以更高或更低抽象级别重复出现。

🌳**对再现性定义的补充说明:** 即使“地球树”可以用于建模抽象过程(如数据结构),或者“智能体”可以通过从“人类”抽象中减去人类特性来定义,但最终都会演化成通用抽象,而非简单地重复利用。

🔗**指针与再现成分的区别:** “指针”是嵌入式智能体在世界模型中复制抽象的过程,而“再现成分”则是自发发生的。例如,人类艺术家对树的理解是指向“树”抽象的指针,而美国市场是指向“市场”抽象的指针。

Published on February 18, 2025 6:04 PM GMT

Consider concepts such as "a vector", "a game-theoretic agent", or "a market". Intuitively, those are "purely theoretical" abstractions: they don't refer to any specific real-world system. Those abstractions would be useful even in universes very different from ours, and reasoning about them doesn't necessarily involve reasoning about our world.

Consider concepts such as "a tree", "my friend Alice", or "human governments". Intuitively, those are "real-world" abstractions. While "a tree" bundles together lots of different trees, and so doesn't refer to any specific tree, it still refers to a specific type of structure found on Earth, and shaped by Earth-in-particular's specific conditions. While tree-like structures can exist in other places in the multiverse, there's an intuitive sense that any such "tree" abstraction would "belong" to the region of the multiverse in which the corresponding trees grow.

Is there a way to formalize this, perhaps in the natural-abstraction framework? To separate the two categories, to find the True Name of "purely theoretical concepts"?


Motivation

Consider a superintelligent agent/optimization process. For it to have disastrous real-world consequences, some component of it would need to reason about the real world. It would need to track where in the world it's embedded, what input-output pathways there are, and how it can exploit these pathways in order hack out of the proverbial box/cause other undesirable consequences.

If we could remove its ability to think about "unapproved" real-world concepts, and make it model itself as not part of the world, then we'd have something plausibly controllable. We'd be able to pose it well-defined problems (in math and engineering, up to whatever level of detail we can specify without exposing it to the real world – which is plenty) and it'd spit out solutions to them, without ever even thinking about causing real-world consequences. The idea of doing this would be literally outside its hypothesis space!

There are tons of loopholes and open problems here, but I think there's promise too.


Ideas

(I encourage you to think about the topic on your own before reading my attempts.)

 

Take 1: Perhaps this is about "referential closure". For concepts such as "vectors" or "agents", we can easily specify the list of formal axioms that would define the frameworks within which these concepts make sense. For things like "trees", however, we would have to refer to the real world directly: to the network of causes and effects entangled with our senses.

... Except that we more or less can, nowadays, specify the mathematical axioms underlying the processes generating our universe (something something Poincaré group). To a sufficiently advanced superintelligence, there'd be no real difference.

Take 2: Perhaps the intuitions are false, and the difference is quantitative, not qualitative.

"Vectors" are concepts such that there's a simple list of axioms under which they're simple to describe/locate: they have low Kolmogorov complexity. By comparison, "trees" have a simple generator, but locating them within that generator's output (the quantum multiverse) takes very many bits.

I guess this is kind of plausible – indeed, it's probably the null hypothesis – but it doesn't feel satisfying.

Especially the pessimistic case: the "continuum" idea doesn't make sense to me. I think there's a big jump between "a human" and "an agent", and I don't see what abstractions could sit between them. (An abstraction over {humans, human governments, human corporations}, which is nevertheless more specific than "an agent in general"? Empirically, humanity hasn't been making use of this abstraction – we don't have a term for it – so it's evidently not convergently useful.)

Take 3: Causality-based definitions. Perhaps "theoretical abstractions" are convergently useful abstractions which can't be changed by any process within our universe (i. e., within the net of causes and effects entangled with our senses)? "Trees" can be wiped out or modified, "vectors" can't be.

This doesn't really work, I think. There are two approaches:

Intuitively, it feels like there's something to the "causality" angle, but I haven't been able to find a useful approach here.

Take 4: Perhaps this is about reoccurrence.

Consider the "global ontology" of convergently useful concepts defined over our universe. A concept such as "an Earthly tree" appears in it exactly once: as an abstraction over all of Earth's trees (which are abstractions over their corresponding bundles-of-atoms which have specific well-defined places, etc.). "An Earthly tree", specifically, doesn't reoccur anywhere else, at higher or lower or sideways abstraction levels.

Conversely, consider "vectors" or "markets". They never show up directly. Rather, they serve as "ingredients" in the makeup of many different "real-world" abstractions. "Markets" can model human behavior in a specific shop, or in the context of a country, and in relation to many different types of "goods" – or even the behavior of biological and even purely physical systems.

Similar for "agents" (animals, humans, corporations, governments), and even more obviously for "vectors".

Potential counterarguments:


Take 4 seems fairly promising to me, overall. Can you spot any major issues with it? Alternatively, a way to more properly flesh it out/formalize it?



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

抽象概念 超智能体 AI安全 理论与现实 再现性
相关文章