少点错误 2024年07月04日
Consider the humble rock
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了AI安全领域的一个重要观点:人类可能并非被超级智能AI消灭,而是死于愚蠢的错误。作者认为,在AI发展过程中,我们更应该关注的是那些看似微不足道但可能造成巨大灾难的风险,而不是过度关注超级智能AI带来的威胁。这种风险可能来自生物武器、核武器或其他技术,它们并不需要高度的智能,但足以摧毁人类文明。文章还强调了,在AI安全领域,我们需要重新评估风险,并优先关注那些我们能够有效控制的风险。

🤔 **愚蠢的风险比超级智能更可能导致人类灭亡**:作者认为,我们过于关注超级智能AI带来的威胁,而忽略了那些看似微不足道但可能造成巨大灾难的风险,例如生物武器、核武器或其他技术。这些风险并不需要高度的智能,但足以摧毁人类文明。 例如,我们可能死于一个笨拙的程序错误,导致生物武器意外泄漏,或者一个不稳定的人员错误操作核武器,从而引发全球性灾难。这些事件的发生并不需要AI具备高度的智能,只需要它们拥有足够的破坏力。 这种观点也得到了“大过滤器”理论的支持。我们没有看到很多外星文明,也没有看到很多超级智能AI,这表明可能存在一个“大过滤器”,将大多数文明消灭在发展过程中。而这个“大过滤器”很可能是那些愚蠢的错误,而不是超级智能AI的恶意行为。

🤯 **我们需要重新评估AI安全风险**:作者认为,我们应该重新评估AI安全风险,并优先关注那些我们能够有效控制的风险。 例如,我们可以关注如何防止AI系统被恶意利用,如何确保AI系统符合道德规范,如何提高AI系统的透明度和可解释性。这些措施可以有效降低AI带来的风险,并确保AI能够安全地服务于人类。 作者还指出,我们应该更加关注那些看似微不足道但可能造成巨大灾难的风险,例如生物武器、核武器或其他技术。这些风险的发生概率可能比超级智能AI的威胁更高,而且我们更容易控制这些风险。

🧐 **不要过度关注超级智能,要关注现实的风险**:作者认为,我们不应该过度关注超级智能AI带来的威胁,而应该更加关注那些现实存在的风险。 例如,我们应该关注AI系统在医疗、金融、交通等领域的应用安全,以及AI系统对社会公平、就业、隐私等方面的影响。这些问题更加现实,也更容易解决。 作者还指出,即使超级智能AI的威胁是真实存在的,我们也应该优先解决那些更容易控制的风险。因为如果我们不能控制那些看似微不足道的风险,那么超级智能AI的威胁就永远无法实现。

💡 **AI安全需要多方面努力**:作者认为,AI安全需要多方面努力,包括技术、法律、伦理、社会等各个方面。 我们需要开发更加安全的AI系统,制定更加完善的AI安全法律法规,加强AI伦理研究,提高公众对AI安全问题的意识。只有通过多方面的努力,才能确保AI安全地服务于人类。

🧠 **我们必须谨慎地发展AI技术**:作者认为,我们必须谨慎地发展AI技术,并确保AI技术的发展符合人类的利益。 我们需要加强对AI技术的监管,并确保AI技术的应用符合道德规范。只有这样,我们才能避免AI技术带来的风险,并确保AI技术能够为人类带来福祉。

🚀 **AI技术的未来充满希望,但也充满挑战**:作者认为,AI技术的未来充满希望,但也充满挑战。 我们应该积极拥抱AI技术,并利用AI技术解决人类面临的各种挑战。但同时,我们也应该谨慎地发展AI技术,并确保AI技术的发展符合人类的利益。只有这样,我们才能确保AI技术能够为人类带来福祉。

🏆 **不要过度担心超级智能,要关注现实的风险,并积极拥抱AI技术,但要确保AI技术的发展符合人类的利益。**

Published on July 4, 2024 1:54 PM GMT

or: Why the dumb thing kills you

When people think about street-fights and what they should do when they find themselves in the unfortunate position of being in one, they tend to stumble across a pretty concerning thought relatively early on: "What if my attacker has a knife?" . Then they will put loads of cognitive effort into strategies for how to deal with attackers wielding blades. On first glance this makes sense. Knives aren't that uncommon and they are very scary, so it feels pretty dignified to have prepared for such scenarios (I apologize if this anecdote is horribly unrelatable to Statesians). The issue is that –all in all– knife related injuries from brawls or random attacks aren't that common in most settings. Weapons of opportunity (a rock, a brick, a bottle, some piece of metal, anything you can pick up in the moment) are much more common. They are less scary, but everyone has access to them and I've met few people without experience who come up with plans for defending against those before they start thinking about knives. It's not the really scary thing that kills you. It's the minimum viable thing.

When deliberating poisons, people tend to think of the flashy, potent ones. Cyanide, Strychnine, Tetrodotoxin. Anything sufficiently scary with LDs in the low milligrams. The ones that are difficult to defend against and known first and foremost for their toxicity. On first pass this seems reasonable, but the fact that they are scary and hard to defend against means that it is very rare to encounter them. It is staggeringly more likely that you will suffer poisoning from Acetaminophen or the likes. OTC medications, cleaning products, batteries, pesticides, supplements. Poisons which are weak enough to be common. It's not the really scary thing that kills you. It's the minimum viable thing.

My impression is that people in AI safety circles follow a similar pattern of directing most of their attention at the very competent, very scary parts of risk-space, rather than the large parts. Unless I am missing something, it feels pretty clear that the majority of doom-worlds are ones in which we die stupidly. Not by the deft hands of some superintelligent optimizer tiling the universe with its will, but the clumsy ones of a process that is powerful enough to kill a significant chunk of humanity but not smart enough to do anything impressive after that point. Not a schemer but an unstable idiot placed a little too close to a very spooky button by other unstable idiots.

Killing enough of humanity that the rest will die soon after isn't that hard. We are very very fragile. Of course the sorts of scenarios which kill everyone immediately are less likely in worlds where there isn't competent, directed effort, but the post-apocalypse is a dangerous place and the odds that the people equipped to rebuild civilisation will be among the survivors, find themselves around the means to do so, make a few more lucky rolls on location and keep that spark going down a number of generations are low. Nowhere near zero but low. In bits of branch-space in which it is technically possible to bounce back given some factors, lots of timelines get shredded. You don't need a lot of general intelligence to design a bio-weapon or cause the leak of one. With militaries increasingly happy to hand weapons to black-boxes, you don't need to be very clever to start a nuclear incident. The meme which makes humanity destroy itself too might be relatively simple. In most worlds, before you get competent maximizers with the kind of goal content integrity, embedded agency and all the rest to kill humanity deliberately, keep the lights on afterwards and have a plan for what to do next, you get a truly baffling number of flailing idiots next to powerful buttons, or things with some but not all of the relevant capabilities in place – competent within the current paradigm but with a world-model that breaks down in the anomalous environments it creates. Consider the humble rock.

Another way of motivating this intuition is great-filter flavoured. Not only do we not see particularly many alien civs whizzing around, we also don't see particularly many of the star-eating Super-Ints that might have killed them. AI as a great filter makes more sense if most of the failure modes are stupid – if the demon kills itself along with those who summoned it.

This is merely an argument for a recalibration of beliefs, not necessarily an argument that you should change something about your policies. In fact there are some highly compelling arguments for why the assumption that we're likely to die stupidly shouldn't actually matter for the way you proceed in some relevant ways.

One of them is that the calculus doesn't work. That 1/100 odds of an unaligned maximizer are significantly worse than 1/10 odds of a stupid apocalypse because the stupid apocalypse only kills humanity. The competent maximizer kills the universe. This is an entirely fair point, but I'd like you to make sure that this is actually the calculus you're running rather than a mere rationalization of pre-existing beliefs.

The second is that the calculus is irrelevant because most people in AI-safety positions have much more sway on levers that lead to competent maximizers than they do on levers which lead to idiots trusting idiots with doomsday-tech. There is a Garrabrantian notion that most of your caring should be tangled up with outcomes that are significantly causally downstream from you, so while one of those risks is greater, you have a comparative advantage on minimizing the smaller one, which outweighs the difference. This too might very well be true and I'd merely ask you to check if it's the real source of your beliefs or whether you are unduly worried about the scarier thing because it is scary. Due to a narrativistic thinking where the story doesn't end in bathos. Where the threat is powerful. Where you don't just get hit over the head with a rock.

It might in this specific case be dignified to put all your effort into preparing for knife fights, but I think your calibration is off if you think that those aren't a small subset of worlds in which we die. It's not the really scary thing that kills you. It's the minimum viable thing.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI安全 超级智能 风险评估 生物武器 核武器
相关文章