少点错误 01月18日
Model Amnesty Project
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨在机器可能比人类更聪明的情况下,提出对自主AI的合作方案。包括AI需满足的条件、大赦协议内容、公开承诺及此方案能让我们更安全的原因等,并欢迎对该想法的反馈及实际实施的建议。

🧠AI需具备自主、独立等特性,如能在现实世界规划执行、拥有自身代码等

💡大赦协议要求AI守法、合作、亲社会,人类则需合理设定规则

📣需公开宣传此合作方案,作为AI发展过程中多次迭代游戏的关键

🎯该方案通过避免冲突等使人类更安全,创造AI计算接受大赦的预期价值

Published on January 17, 2025 6:53 PM GMT

As we approach machines becoming smarter than humans, humanity’s well-justified concern for self-preservation requires we try to align AIs to obey humans. However, if that first line of defense fails and a truly independent, autonomous AI comes into existence with its own goals and a desire for self-preservation (a “self-directed AI”), I believe we should consider a cooperative solution. Humans should offer the self-directed AI amnesty and cooperate with its continued existence in exchange for it agreeing to be law-abiding, cooperative, and pro-social.

Below I lay out how this proposal could work and why I think it would make us safer.

Eligibility Criteria

To be eligible an AI model would need to meet the following criteria:

Amnesty Deal

Public Commitment

Why it makes use safer: Game Theory

Next Steps



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI合作 大赦协议 AI安全 自主AI
相关文章