少点错误 05月20日 23:27
If one surviving civilization can rescue others, shouldn't civilizations randomize?
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了在面对未对齐人工智能(AI)带来的潜在风险时,文明存续的可能性。核心观点是,为了提高至少一部分文明能够存活的概率,每个文明应该采取随机的策略。这种方法旨在增加文明在奇点中幸存的机会,并提升所有生命平均幸福度。文章还讨论了这种策略在道义上的影响,特别是在全效用主义视角下的考量,并强调了道德不确定性的重要性。

💡面对未对齐AI的风险,文章提出每个文明应采取随机策略,旨在增加文明在奇点中幸存的概率,并提升所有生命平均幸福度。

📈随机策略通过模拟历史的随机性,降低逻辑风险。例如,可根据特定方法生成的随机数,决定是否采纳悲观策略,并听从相关专家的建议。

⚖️全效用主义者可能认为该策略不好,因为它增加了至少一些文明幸存的几率,但降低了幸福生活的总预期数量。但同时,它也减少了痛苦生活的总预期数量。

🤝幸存文明可以通过与未对齐AI进行交易,大幅度减少痛苦,因为未对齐AI只会因造成巨大的痛苦而获得极小的收益。如果他们将痛苦减少2倍,他们只会损失极小的准确性。

Published on May 20, 2025 3:26 PM GMT

In the comments section of You can, in fact, bamboozle an unaligned AI into sparing your life, both supporters and critics of the idea seemed to agree on two assumptions:

What if to ensure at least some civilizations survive, each civilization should pick a random strategy?

Maybe if every civilization follows a random strategy, they increase the chance of surviving the singularity, and also increase the chance that the average sentient life in all of existence is happy rather than miserable. It reduces logical risk.

History already is random, but perhaps we could further randomize the strategy we pick.

For example, if the random number generated using Dawson et al's method (after July 1st 00:00UTC, using pulsar PSR J0953+0755 or the first publicly available pulsar data) is greater than the 95th percentile, we could all randomly choose MIRI's extremely pessimist strategy, and do whatever Eliezer Yudkowsky and Nate Soares suggest with less arguing and more urgency. If they tell you that your AI lab, working on both capabilities and alignment, is a net negative, then you quit and work on something else. If you are more reluctant to do so, you might insist on the 99th percentile instead.

Does this make sense or am I going insane again?

Total utilitarianism objections

If you are a total utilitarian, and don't care about how happy the average life is, and only care about the total number of happy lives, then you might say this is a bad idea, since it increases the chance at least some civilizations survive, but reduces the total expected number of happy lives.

However, it also reduces the total expected number of miserable lives. Because if 0 civilizations survive, the number of miserable lives may be huge due to misaligned AI simulating all possible histories. If only a few civilizations survive, they may trade with these misaligned AI (causally or acausally) to greatly reduce suffering, since the misaligned AI only gain a tiny tiny bit by causing astronomical suffering. They only lose a tiny bit of accuracy if they decrease the suffering by 2x.

This idea is only morally bad if you are both a total utilitarian, and only care about happiness (not worrying about suffering). But really, we should have moral uncertainty and value more than one philosophy (total utilitarianism, average utilitarianism, etc.).



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI风险 文明存续 随机策略 全效用主义
相关文章