少点错误 2024年07月16日
[Research log] The board of Alphabet would stop DeepMind to save the world
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了AI安全社区的一种观点,即前沿AI实验室在开发变革性AI的过程中,是否需要政府干预来阻止这场“自杀竞赛”。作者认为,如果实验室决策者意识到风险巨大,他们可能会选择退出竞赛,而不是一味追求AGI的开发。

🤖【风险认知】前沿AI实验室的决策者如果认识到继续竞赛可能导致人类毁灭,他们可能会选择停止开发,即使这意味着失去竞争优势。

🔍【决策场景】文章设想了一个场景,Alphabet董事会面对高风险的AGI开发,可能会选择放弃竞赛,转而专注于安全措施的提升。

🌍【政府干预】作者质疑是否真的需要政府施加外部压力来改变实验室的激励结构,认为实验室自身可能基于风险考量作出合理决策。

🧩【决策因素】文章分析了影响决策的几个因素,如信息不透明、时间压力、关键考量的缺失等,这些因素可能导致决策者无法作出最佳选择。

🔬【风险模型】由于AGI开发是前所未有的事件,用于评估风险的模型可能存在误差,导致决策失误。

Published on July 16, 2024 4:59 AM GMT

Produced as part of the ML Alignment & Theory Scholars Program - Summer 2024 Cohort.

This post is not a finished research. I’m unconfident in the claims I’m making here, but I thought putting it out there for feedback would help me decide what to focus on next in the program.

Are government interventions necessary to stop the suicide race?

The zeitgeist I got from the AI Safety community since I joined, seems to accept as fact that Frontier AI Labs are locked knowingly in a suicidal race towards developing transformative AI, and that any solution will need to involve strong external pressure to stop them, either in the form of an international coalition imposing regulations which shift the incentives of Labs, or even more desperate measure like doing a pivotal act.

From this point of view, it seems that AI risk is mostly driven by game theory. The economic and personal incentives faced by the stakeholders of each Frontier AI Lab determine their actions, and they will proceed this way until AGI is developed, or until a sufficiently large external force changes the incentive landscape sufficiently. Therefore, the only way to make sure Labs don’t gamble the future of the world when building an AGI is to convince governments to implement policies which shift those incentives.

I now believe that this view is wrong, and that Frontier AI Labs would get out of the race if they thought the risks were sufficiently large and the consequences sufficiently dire for the world and for themselves.

Claim: If every decision maker in a Frontier AI Lab thought they were in a suicide race, and that their next development would bring with near certainty the destruction of humanity, they would decide to leave the “AGI at all costs” race, no matter the actions of other actors.

Below, I present a scenario which I find plausible, in which a Frontier Lab decides to drop out of the race because of this.

Scenario: The board of Alphabet decides to stop the race

Disclaimer: Handwaving a lot of details.

Situation: Frontier AI Labs are still locked in a race to be the first to develop AGI. It is widely believed that the coming generation of models might pass the threshold. We are in crunch time. A previous unsuccessful attempt by a model to take control of key infrastructure makes the possibility of X-risk clear in everyone’s mind. The team at Google DeepMind is hard at work preparing the next training run, which they believe will be the last one.

Demis Hassabis calls for a board meeting of Alphabet, where he present the current tactical situation. All board members get convinced of the following claims:

After having received this information, the board convenes to decide how to proceed. They consider two decisions:

I expect that, in this situation, the executives of Alphabet would decide to drop out of the race, as continuing would be such a high probability of death to everyone, including the board members themselves.

Would they press the button?

I see three possible reasons why the board might want to proceed despite the risks:

I think all of those are unlikely. Am I missing some other reason why they would do it?

Why real life will be much more risky than this scenario

Even if in this scenario the board of Alphabet can be reasonably believed to take the right call and stop development, I expect that such a clear-cut vision of the consequences will never be available. I expect that various forms of imperfect information and changes to the payoff matrix will make it less likely that they would drop out of the race before it’s too late.

However, I’m interested in knowing what exactly are the factors which prevent such an ideal scenario from happening, as it could inform my priorities for reducing AI risks. I’m specifically interested in which factors prevent decision makers from having such a complete view of the situation, and which interventions beside policy could improve those decisions.

A short list of factors which I expect to cause decision-making to be less than ideal:

  1. ^


Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI安全 政府干预 决策风险 AGI开发
相关文章