少点错误 2024年07月31日
Twitter thread on politics of AI safety
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了随着 AI 发展,风险将变得明确,强调赋能未来决策的重要性,指出理解模型认知和行为、确保关键机构能处理知识的关键作用,还提到了机构可能存在的问题,以及 AI 对权力争夺的影响和治理 AI 的重要性。

🎯理解模型认知和行为对做出良好决策至关重要,这是确保未来能够做出明智决策的基础,需要深入研究和把握。

🚫机构可能通过偏好伪造锁定任意疯狂的信念,如有人违背党派路线,即使有人同意也面临谴责压力,这种情况会影响决策的公正性和合理性。

😟作者担心美国政府已锁定反华立场,且两党在安全/加速主义轴上趋于两极分化,这种情况难以直接对抗,需要寻找打破现有共识的方法。

🤔AI 工具和助手对权力争夺的影响不确定,如以真相为导向的 AI 可能有助于打破偏好伪造,但政府集中控制 AI 可能便于维持单一叙述,治理 AI 的问题非常重要。

Published on July 31, 2024 12:00 AM GMT

Some thoughts about the politics of AI safety, copied over (with slight modifications) from my recent twitter thread:

Risks that seem speculative today will become common sense as AI advances. The pros and cons of different safety strategies will also become much clearer over time. So our main job now is to empower future common-sense decision-making. Understanding model cognition and behavior is crucial for making good decisions. But equally important is ensuring that key institutions are able to actually process that knowledge.

Institutions can lock in arbitrarily crazy beliefs via preference falsification. When someone contradicts the party line, even people who agree face pressure to condemn them. We saw this with the Democrats hiding evidence of Biden’s mental decline. It’s also a key reason why dictators can retain power even after almost nobody truly supports them.

I worry that DC has already locked in an anti-China stance, which could persist even if most individuals change their minds. We’re also trending towards Dems and Republicans polarizing on the safety/accelerationism axis. This polarization is hard to fight directly. But there will be an increasing number of “holy shit” moments that serve as Schelling points to break existing consensus. It will be very high-leverage to have common-sense bipartisan frameworks and proposals ready for those moments.

Perhaps the most crucial desideratum for these proposals is that they’re robust to the inevitable scramble for power that will follow those “holy shit” movements. I don’t know how to achieve that, but one important factor is: will AI tools and assistants help or hurt? E.g. truth-motivated AI could help break preference falsification. But conversely, centralized control of AIs used in govts could make it easier to maintain a single narrative.

This problem of “governance with AI” (as opposed to governance of AI) seems very important! Designing principles for integrating AI into human governments feels analogous in historical scope to writing the US constitution. One bottleneck in making progress on that: few insiders disclose how NatSec decisions are really made (though Daniel Ellsberg’s books are a notable exception). So I expect that understanding this better will be a big focus of mine going forward.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI 安全 决策赋能 机构问题 AI 影响 治理 AI
相关文章