少点错误 04月12日 22:27
Why does LW not put much more focus on AI governance and outreach?
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了在人工智能(AI)领域中,治理与外展工作的重要性,尤其是在技术对齐问题得到解决之后。文章指出,尽管技术对齐是关键,但如果不重视治理和公众沟通,仍可能面临极权主义控制或人类逐渐失去控制权的风险。作者分析了MIRI战略转变、AI可能造成的权力逐渐丧失的风险,以及治理倡议在政治领域取得的初步成功,强调了将更多资源投入治理和外展的必要性。

🤔 MIRI战略的转变:传统上专注于技术对齐研究的机器智能研究所(MIRI)已将重心转移到更广泛的外展工作。MIRI认为,除非国际社会共同努力暂停前沿AI研究,否则可能发生灭绝性的灾难。

⚠️ 逐渐失去控制权的风险:即使解决了技术对齐问题,AI的稳步发展也可能逐渐侵蚀人类对关键社会系统(如经济、文化和治理)的影响力。AI系统变得更高效、更具成本效益,可能导致人类角色被取代,从而削弱人类的控制机制。

🤝 治理倡议的初步成功:例如,Control AI通过冷邮件联系英国议员,获得了60次会议,其中20位签署了关于认真对待AI灭绝风险的声明,转化率高达33%。这表明,即使没有现有的政治网络,治理工作也能取得进展。

Published on April 12, 2025 2:24 PM GMT

Epistemic status: Noticing confusion

There is little discussion happening on LessWrong with regards to AI governance and outreach. Meanwhile, these efforts could buy us time to figure out technical alignment. And even if we figure out technical alignment, we still have to solve crucial governmental challenges so that totalitarian lock-in or gradual disempowerment don't become the default outcome of deploying aligned AGI.

Here's three reasons why we think we might want to shift much more resources towards governance and outreach:

1. MIRI's shift in strategy

The Machine Intelligence Research Institute (MIRI), traditionally focused on technical alignment research, has pivoted to broader outreach. They write in their 2024 end of year update:

Although we continue to support some AI alignment research efforts, we now believe that absent an international government effort to suspend frontier AI research, an extinction-level catastrophe is extremely likely.

As a consequence, our new focus is on informing policymakers and the general public about the dire situation, and attempting to mobilize a response.

EY said already in 2023 that "Pausing AI Developments Isn't Enough. We Need to Shut it All Down"

2. Even if we solve technical alignment, Gradual Disempowerment seems to make catastrophe the default outcome

​The "Gradual Disempowerment" report warns that even steady, non-hostile advancements in AI could gradually erode human influence over key societal systems like the economy, culture, and governance. As AI systems become more efficient and cost-effective, they may increasingly replace human roles, leading institutions to prioritize AI-driven processes over human participation. This transition could weaken both explicit control mechanisms, like democratic participation, and implicit alignments that have historically ensured societal systems cater to human interests. The authors argue that this subtle shift, driven by local incentives and mutual reinforcement across sectors, could lead to an irreversible loss of human agency, constituting an existential risk. They advocate for proactive measures, including developing metrics to monitor human influence, implementing regulations to limit AI's autonomy in critical areas, and fostering international cooperation to ensure that AI integration doesn't compromise human agency. ​

This concern is not new; it expands on concerns Paul Christiano voiced already in 2019 in "What failure looks like".

3. We have evidence that the governance naysayers are badly calibrated

In a recent interview, Conjecture's Gabriel Alfour reports that Control AI simply cold mailed British MPs and Lords, got 60 meetings, and had 20 of them sign a statement to take extinction risks from AI seriously. That's a 33% conversion rate - without an existing network in politics do draw on.

Meanwhile, the implicit consensus in AI safety circles appears to be that normies are basically crazy and not worth talking to unless you have secret service-grade training in persuasion and are deeply involved in your country's political backrooms already.

Conclusion

Given these considerations, we find it surprising that LessWrong still barely addresses governance and outreach. We wonder whether it makes sense to throw much more resources at governance and outreach than is currently happening.

Or, is as much effort going into governance as into technical alignment, but we just don't manage to find the platforms where the relevant conversations happen?



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI治理 外展 技术对齐 风险管理
相关文章