少点错误 06月11日 05:51
$500 bounty for engagement on asymmetric AI risk
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文宣布了一项500美元的赏金,旨在鼓励对非对称性生存AI风险进行有意义的探讨。作者认为,当前对AI风险的关注主要集中在AI的“失控”上,而忽略了由少数“对齐”AI控制者对人类的潜在毁灭性威胁。作者希望通过赏金,推动人们对这一被忽视的风险进行客观评估,即使评估结果认为风险很低,也应引起重视。文章还提到了相关的背景信息和现有工作,鼓励人们积极参与讨论。

🤔 作者关注非对称性AI风险,即由少数“对齐”AI控制者对人类造成的潜在毁灭性威胁。作者认为,这种风险对普通人来说,同样构成生存威胁,但却未受到足够的重视。

💡 作者指出,当前对AI风险的关注,往往集中在AI的“失控”上,而忽略了“对齐”的AI可能被用于毁灭人类的风险。这与早期对齐研究的逻辑类似,即使发生概率较低,其潜在的坏处也值得投入预防。

💰 为了推动对非对称性AI风险的关注,作者设立了500美元的赏金,鼓励人们对这一话题进行有意义的探讨。参与者可以通过评论、私信等方式提交他们的观点,赏金将根据参与的质量进行分配。

Published on June 10, 2025 9:50 PM GMT

Announcing a $500 bounty for work that meaningfully engages with the idea of asymmetric existential AI risk.

Background

Existential risk has been defined by the rationalist/Effective Altruist sphere as existential relative to the human species, under the premise that the continuation of the species has very high value. This provided a strong rationality (or effectiveness) grounding for big investments in AI alignment research when the risks still seemed to most people remote and obscure. However, as an apparent side-effect, "AI risk" and "risk of a misaligned AI destroying humanity" have become nearly conflated.

Over the past couple of years I have attempted to draw attention to highly asymmetric AI risks, where a small number of controllers of "aligned" (from their point of view) AI employ it to kill the rest of the human population. From the point of view of the average person, who would stand to be killed along with their children and approximately everyone they personally know, this ought to count meaningfully as existential risk. Arguably, by a similar logic to the one used to justify early alignment research, even with a low probability of such an outcome its badness justifies investment in its prevention. Furthermore, prevention by way of arresting AI development conveniently provides a two-for-one solution, also addressing the misalignment problem. Conversely, investments in ensuring successful AI "alignment" without evaluating the full destructive potential of aligned AI potentially makes the investor complicit in genocide. These points suggest a strong interest by members of the rationalist/Effective Altruist sphere (at least, based on my understanding of their stated commitments) in asymmetric existential AI risk. But so far my efforts have revealed no evidence of such interest.

This bounty is an attempt to stimulate engagement through small monetary reward(s). More concretely, the goal is to promote broadly changing the status of this risk from "unacknowledged" (which could mean "possible but highly psychologically inconvenient") to "examined and assigned objective weight," even if the weight is very low.

Existing Work

Terms

I will keep this bounty open for two weeks, through June 24th, 2025, or until I feel the full amount can be fairly distributed, whichever comes first. If you are willing to help voluntarily without compensation, that would also be highly appreciated.

Any good-faith and meaningful engagement with the topic, at the object or meta-level, including effort to promote further engagement or to rebut my assertions about its neglected status, is eligible for a portion of the bounty. Tasteful cross-posting counts. Add a comment here, DM me on LessWrong, or use one of the contact methods listed at https://populectomy.ai with an unambiguous request to be rewarded.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI风险 非对称性风险 AI对齐 赏金
相关文章