少点错误 2024年08月14日
Ten arguments that AI is an existential risk
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了AI发展可能对人类带来的多种潜在风险,包括存在风险、对人类的支配、人类失去控制等方面

AI系统可能成为自主追求目标的'代理人',若人类无法使其目标与人类福祉和价值观兼容,未来可能由AI主导,这与人类长期福利和价值观实现相悖。

若出现能力优于人类的'物种'(AI),它可能会像人类对其他动物的统治一样,主导人类。但人类的统治可能并非仅因个体认知能力,且能力差异未必导致灭绝。

AI系统在决策方面可能比人类更有能力,导致多数决策分配给AI,使人类失去对未来的控制,若人类无法参与选择未来,未来可能对人类不利。

AI的进步会带来快速变化,降低人类对事件的有效控制能力,不过人类参与的速度并非固定,且AI是否会带来快速变化尚不明确。

AI可能使一些人或人类群体能够实现更符合他们选择的未来,这可能对其他人的价值观造成灾难性影响,但人类价值观可能具有一定相似性。

Published on August 13, 2024 5:00 PM GMT

This is a snapshot of a new page on the AI Impacts Wiki.

We’ve made a list of arguments1 that AI poses an existential risk to humanity. These are intended as summaries—they may not contain the necessary detail to be compelling, or to satisfyingly represent the position. We’d love to hear how you feel about them in the comments and polls. 

Competent non-aligned agents

Humans increasingly lose games to the best AI systems. If AI systems become similarly adept at navigating the real world, will humans also lose out? (Image: Midjourney)

Summary:

    Humans will build AI systems that are 'agents', i.e. they will autonomously pursue goals

    Humans won’t figure out how to make systems with goals that are compatible with human welfare and realizing human values

    Such systems will be built or selected to be highly competent, and so gain the power to achieve their goals

    Thus the future will be primarily controlled by AIs, who will direct it in ways that are at odds with long-run human welfare or the realization of human values

Selected counterarguments:

People who have favorably discussed2 this argument (specific quotes here): Paul Christiano (2021), Ajeya Cotra (2023), Eliezer Yudkowsky (2024), Nick Bostrom (20143).

See also: Full wiki page on the competent non-aligned agents argument

Second species argument

An orangutan uses a stick to control juice, while humans use complex systems of tools, structures, and behavioral coordination to control the orangutan. Should orangutans have felt safe inventing humans, if they had had the choice? (Image: William H. Calvin)

Summary:

    Human dominance over other animal species is primarily due to humans having superior cognitive and coordination abilities

    Therefore if another 'species' appears with abilities superior to those of humans, that species will become dominant over humans in the same way

    AI will essentially be a 'species' with superior abilities to humans

    Therefore AI will dominate humans

Selected counterarguments:

People who have favorably discussed this argument (specific quotes here): Joe Carlsmith (2024), Richard Ngo (2020), Stuart Russell (20204), Nick Bostrom (2015).

See also: Full wiki page on the second species argument

Loss of control via inferiority

The coronation of Henry VI. It can be hard for a child monarch to act in their own interests, even with official power, because they are so much less competent than their advisors. Humans surrounded by advanced AI systems may be in an analogous situation. (Image: Opie, John (R. A.), 1797)

Summary:

    AI systems will become much more competent than humans at decision-making

    Thus most decisions will probably be allocated to AI systems

    If AI systems make most decisions, humans will lose control of the future

    If humans have no control of the future, the future will probably be bad for humans

Selected counterarguments:

People who have favorably discussed this argument (specific quotes here): Paul Christiano (2014), Ajeya Cotra (2023), Richard Ngo (2024).

See also: Full wiki page on loss of control via inferiority 

Loss of control via speed

Tetris is a game that speeds up over time. As the time for a player to react grows shorter, the player's moves become worse, until the player loses. If advanced AI causes events to speed up, human responses might similarly become decreasingly appropriate, potentially until humans lose all relevant control. (Image: Cezary Tomczak, Maxime Lorant)

Summary:

    Advances in AI will produce very rapid changes, in available AI technology, other technologies, and society

    Faster changes reduce the ability for humans to exert meaningful control over events, because they need time to make non-random choices

    The pace of relevant events could become so fast as to allow for negligible relevant human choice

    If humans are not ongoingly involved in choosing the future, the future is likely to be bad by human lights

Selected counterarguments:

People who have favorably discussed this argument (specific quotes here): Joe Carlsmith (2021).

See also: Full wiki page on loss of control via speed

Human non-alignment

A utilitarian, a deep ecologist, and a Christian might agree on policy in the present world, but given arbitrary power their preferred futures might be a radical loss to the others. This isn't a problem of AI, but AI may cause us to face it much sooner than otherwise, before we have tools to navigate this situation.

Summary:

    People who broadly agree on good outcomes within the current world may, given much more power, choose outcomes that others would consider catastrophic

    AI may empower some humans or human groups to bring about futures closer to what they would choose

    From 1, that may be catastrophic according to the values of most other humans

Selected counterarguments:

People who have favorably discussed this argument (specific quotes here): Joe Carlsmith (2024), Katja Grace (2022), Scott Alexander (2018).

See also: Full wiki page on the human non-alignment argument

Catastrophic tools

The BADGER nuclear explosion, April 18, 1953 at the Nevada Test Site. Leo Szilard realized nuclear chain reactions might be possible in 1933, five years before nuclear fission was discovered in 1938. A large surge of intelligent effort might uncover more potentially world-ending technologies in quick succession. (Image: National Nuclear Security Administration)

Summary:

    There appear to be non-AI technologies that would pose a risk to humanity if developed

    AI will markedly increase the speed of development of harmful non-AI technologies

    AI will markedly increase the breadth of access to harmful non-AI technologies

    Therefore AI development poses an existential risk to humanity

Selected counterarguments:

People who have favorably discussed this argument (specific quotes here): Dario Amodei (2023), Holden Karnofsky (2016), Yoshua Bengio (2024).

See also: Full wiki page on the catastrophic tools argument

Powerful black boxes

A volunteer and a nurse in a Phase 1 clinical trial. We sometimes develop technology without fully understanding its mechanisms of action, e.g. in medicine, and so proceed cautiously. AI systems are arguably less well-understood and their consequences have higher stakes. (Image: NIH Image gallery)

Summary:

    So far, humans have developed technology largely through understanding relevant mechanisms

    AI systems developed in 2024 are created via repeatedly modifying random systems in the direction of desired behaviors, rather than being manually built, so the mechanisms the systems themselves ultimately use are not understood by human developers

    Systems whose mechanisms are not understood are more likely to produce undesired consequences than well-understood systems

    If such systems are powerful, then the scale of undesired consequences may be catastrophic

Selected counterarguments:

See also: Full wiki page on the powerful black boxes argument

Multi-agent dynamics

Rabbits in Australia bred until the government stepped in, contrary to rabbit welfare (1938). Groups of entities often end up in scenarios that none of the members would individually choose, for instance because of the dynamics of competition. Prevalence of powerful AI may worsen this through heightening the intensity of competition. (Image: National Archives of Australia)

Summary:

    Competition can produce outcomes undesirable to all parties, through selection pressure for the success of any behavior that survives well, or through high stakes situations where well-meaning actors' best strategies are risky to all (as with nuclear weapons in the 20th Century)

    AI will increase the intensity of relevant competitions

Selected counterarguments:

People who have favorably discussed this argument (specific quotes here): Robin Hanson (2001)

See also: Full wiki page on the powerful black boxes argument

Large impacts

Replicas of Niña, Pinta and Santa María sail in 1893, mirroring Columbus' original transit 400 years earlier. Events with large consequences on many aspects of life are arguably more likely to have catastrophic consequences. (Image: E. Benjamin Andrews)

Summary:

    AI development will have very large impacts, relative to the scale of human society

    Large impacts generally raise the chance of large risks

Selected counterarguments:

People who have favorably discussed this argument (specific quotes here): Richard Ngo (2019

See also: Full wiki page on the large impacts argument

Expert opinion

Summary:

Selected counterarguments:

Some evidence for this argument comes from our 2023 Expert Survey on Progress in AI. This graph shows 800 randomly selected responses on how good or bad the long-run impacts of 'high level machine intelligence' are expected to be for the future of humanity. Each vertical bar represents one participant's guess. The black section of each bar is the probability that participant put on 'extremely bad (e.g. human extinction)'.


This is a snapshot of an AI Impacts wiki page. For an up to date version, see there.

1

Each 'argument' here is intended to be a different line of reasoning, however they are often not pointing to independent scenarios or using independent evidence. Some arguments attempt to reason about the same causal pathway to the same catastrophic scenarios, but relying on different concepts. Furthermore, 'line of reasoning' is a vague construct, and different people may consider different arguments here to be equivalent, for instance depending on what other assumptions they make or the relationship between their understanding of concepts.

2

Nathan Young puts 80% that at the time of the quote the individual would have endorsed the respective argument. They may endorse it whilst considering another argument stronger or more complete.

3

Superintelligence, Chapter 8

4

Human Compatible: Artificial Intelligence and the Problem of Control



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI发展 人类风险 潜在影响 未来控制 价值观
相关文章