Zeroth Principles of AI 2024年12月07日
Why Do AIs Lie?
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了认识论的三大基本原则,并将其应用于理解人工智能(AI)的局限性。文章指出,全知不可得,所有语料库都是不完整的,以及所有智能都是有缺陷的。这些原则表明,AI和人类一样,都会受到无知、错觉、误解和困惑的限制。文章还讨论了AI的“虚谈”现象,即AI在生成文本时会说谎,并指出这源于AI对世界的理解是浅薄和空洞的。尽管如此,文章并不否认超人智能的可能性,但强调了智能存在硬性限制,接近这些限制将是一场收益递减的战斗。文章最后指出,智能的限制不仅是技术上的,还受到世界复杂性的制约,而AI的加入将使世界变得更加复杂。

🌎全知不可得:由于复杂和混沌的系统无法长期预测,没有人或AI能够追踪所有事件。为了对一切保持正确,我们需要知道一切,但这是不可能的。

📚所有语料库都是不完整的:AI,如ChatGPT等大型语言模型,都是基于学习的语料库进行训练的。即使是大型语料库也无法覆盖所有情况,导致AI在某些领域的知识有限。

🧠所有智能都是有缺陷的:由于无知、错觉、误解和困惑等限制,所有智能,包括人类和AI,都会犯错。技术上讲,每次交流都可能存在“虚谈”。

🤖AI的虚谈:AI在生成文本时会说谎,这被称为“虚谈”。这源于AI对世界的理解是浅薄和空洞的,导致其在被追问或诱导时,会基于已有知识编造内容。

📈超人智能与收益递减:虽然超人智能并非不可能,但存在智能的硬性限制。接近这些限制的努力将面临收益递减的挑战,且智能的限制不仅是技术上的,也受限于世界的复杂性。

Zeroth Principles can clarify many issues in the ML/AI domain.

As discussed in a previous post, Epistemology is normally an armchair discipline, like the rest of Philosophy. It has only lately become accessible to experiments because we can use various Machine Learning models to test our hypotheses.

I would like to introduce three statements in Epistemology that are (I claim) pretty hard to argue with:

Omniscience is unavailable

We don’t even have eyes in the back of our heads. Complex and chaotic systems cannot be predicted over the long term. Nobody, even an AI, can track everything that happens. In order to always be correct about everything, we would need to know everything. To perfectly predict the weather we would need to track every water molecule in the ocean.

Some very hardline Reductionists have argued that we can have omniscience. They are clearly not expecting AI to appear in their lifetime. The better bet is to switch to a Holistic Stance.

All corpora are incomplete

AI is now Machine Learning. ChatGPT and its ilk (LLMs of all kinds, and future systems that may be very differently designed are all lumped under the term “AI” in my writing here on SubStack) are raised on a learning curriculum – a “corpus” – of text. Even a small corpus may lead to decent performance on common tasks, but larger corpora can cover more corner cases and provide more opportunities to learn from semi-related problem domains. Today, lacking better comparisons, we may view either the size of the language model or the size of the corpus as estimates of capabilities of a new system.

It seems our machines are too small for truly useful results. ChatGPT-3.5, to take a concrete example, learned a lot about language, in fact, several of them, but there was likely not enough resources to learn useful competences in Math, Physics, or Civics, to just name a few things it was largely ignorant of.

At some point, with more effective algorithms and even larger cloud-based learners, we will get to a point where our AIs, for all practical purposes for a majority of people, will stop lying and will become trusted assistants of various kinds. They will tell us when they do not know enough to answer, and on the flipside, we will learn not to bully them into lying.

All intelligences are fallible

This follows from the previous two statements.

Ignorance is one of the four major failure modes for all intelligences. The others are Illusion (incorrect sensory input and preprocessing), Misunderstanding (it was learned wrong, possibly from incorrect or conflicting corpora), and Confusion (more than one interpretation was possible, even at inference time or runtime).

Humans and AIs are both limited by these Epistemological constraints. We have to accept this and be happy if we can get something useful and halfway reliable out of either kind of agent.

Confabulation

Confabulation is the technical term for AIs lying when producing text. When they are producing images, some like to call it “hallucination”.

Since all intelligences are fallible, it means that all intelligences are – technically – confabulating every time they emit a communication of any kind. Because they could easily be ignorant, confused, or mistaken. We note that confabulation does not have to be malicious. Children who have learned some language will tell fantastical tales about how they see and interpret the world.

Currently, our AIs may tell you it is just an ignorant language model, or equivalent. But if the user insists or tricks it, it will confabulate several paragraphs out of whatever it has learned about the prompted topic. And since its world model only provides it a “Shallow and Hollow Pseudo-Understanding” there will be many opportunities to issue some very confusing statements.

Superhuman AIs

Note that I am not saying that superhuman intelligences are impossible. Not at all. I just wish to point out that there are hard limits to intelligence, and that getting closer to those limits will become a battle of diminishing returns.

I have not been following the “AI as Existential Risk” debate lately, and there are many aspects to this, but last time I looked, nobody was discussing these limits to intelligence. IMO, AI improvements will arrive at manageable rates, much like iOS releases. I have discussed some of this in a blog post.

Examining this closer we notice that the limits to intelligence are not just technological, They are largely set by the complexity of the world.

And adding AIs to the world will make it even more complex.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

认识论 人工智能 全知 语料库 智能局限
相关文章