少点错误 02月22日
The Sorry State of AI X-Risk Advocacy, and Thoughts on Doing Better
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文指出当前AI风险宣传的困境,认为对ML研究员、政策制定者和网络用户的说服效果已达瓶颈。文章强调,应将目标转向普通大众,因为他们对AI风险的认知度较低,但潜在政治影响力巨大。通过提高大众对AI风险的认知,可以形成对政府的政治压力,促使其重视并解决相关问题。文章建议利用喜剧演员、报纸、播客等渠道,以通俗易懂的方式传播AI风险知识,引发大众对“谁将控制ASI”等问题的思考,从而推动更广泛的社会关注和行动。

🎯 **目标受众转变**: 当前AI风险宣传应从ML研究员、政策制定者和网络用户转向普通大众,因为前者已形成固定观念,而后者认知度低但潜力巨大。

📢 **大众认知提升**: 普通大众对AI的理解多停留在聊天机器人层面,缺乏对硬着陆、AI加速AI研究等概念的认知。因此,需要用通俗易懂的方式提高他们对AI风险的认知。

🎭 **宣传渠道多样化**: 可利用喜剧演员、报纸、播客等渠道,以更贴近大众的方式传播AI风险知识,引发更广泛的社会关注。

🤔 **引发深度思考**: 引导大众思考“谁将控制ASI”等问题,激发他们对政府、企业等机构的潜在不信任感,从而推动对AI风险的重视。

Published on February 21, 2025 8:15 PM GMT

First, let me quote my previous ancient post on the topic:

Effective Strategies for Changing Public Opinion

The titular paper is very relevant here. I'll summarize a few points.

    The main two forms of intervention are persuasion and framing.Persuasion is, to wit, an attempt to change someone's set of beliefs, either by introducing new ones or by changing existing ones.Framing is a more subtle form: an attempt to change the relative weights of someone's beliefs, by empathizing different aspects of the situation, recontextualizing it.There's a dichotomy between the two. Persuasion is found to be very ineffective if used on someone with high domain knowledge. Framing-style arguments, on the other hand, are more effective the more the recipient knows about the topic.Thus, persuasion is better used on non-specialists, and it's most advantageous the first time it's used. If someone tries it and fails, they raise the recipient's domain knowledge, and the second persuasion attempt would be correspondingly hampered. Cached thoughts are also in effect.Framing, conversely, is better for specialists.

My sense is that, up to this point, AI risk advocacy targeted the following groups of people:

Persuasion

I think all of the above demographics aren't worth trying to persuade further at this point in time. It was very productive before, when they didn't yet have high domain knowledge related to AI Risk specifically, and there's been some major wins.

But further work in this space (and therefore work on all corresponding advocacy methods, yes) is likely to have ~no value.

Among those groups, we've already convinced ~everyone we were ever going to convince. That work was valuable and high-impact, but the remnants aren't going to budge in response to any evidence short of a megadeath AI catastrophe.[1]

Hell, I am 100% behind the AI X-risk being real, and even I'm getting nauseated at how tone-deaf, irrelevant, and impotent the arguments for it sound nowadays, in the spaces in which we keep trying to make them.

 

A Better Target Demographic

Here's whom we actually should be trying to convince inform: normal people. The General Public.

If we can raise the awareness of the AGI Doom among the actual general public (again, not the small demographic of terminally online people), that will create significant political pressure on the USG, giving politicians an incentive to have platforms addressing the risks.

The only question is how to do that. I don't have a solid roadmap here. But it's not by writing viral LW/Xitter blog posts.

Some scattershot thoughts:

Overall, I expect that there's a ton of low-hanging high-impact fruits in this space, and even more high-impact clever interventions that are possible (in the vein of harfe's idea).

 

Extant Projects in This Space?

Some relevant ones I've heard about:

 

Framing

Technically, I think there might be some hope for appealing to researchers/academics/politicians/the terminally online, by reframing the AI Risk concerns in terms they would like more.

All the talk about "safety" and "pauses" have led to us being easy to misinterpret as unambitious, technology-concerned, risk-averse luddites. That's of course incorrect. I, at least, am 100% onboard with enslaving god, becoming immortal, merging with the machines, eating the galaxies, perverting the natural order to usher in an unprecedented age of prosperity, forcing the wheels of time into reverse to bring the dead back to life, and all that good stuff. I am pretty sure most of us are like this (if perhaps not in those exact terms).

The only reason I/we are not accelerationists is because the current direction of AI progress is not, in fact, on the track to lead us to that glorious future. It's instead on the track to get us all killed like losers.

So a more effective communication posture might be to empathize this: frame the current AI paradigm as a low-status sucker's game, and suggest alternative avenues for grabbing power. Uploads, superbabies, adult intelligence enhancement, more transparent/Agent Foundations-y AI research, etc. Reframing "AI Safety" as being about high-fidelity AI Control might also be useful. (It's mostly about making AIs Do What You Mean, after all, and the best alignment work is almost always dual-use.)

If the current paradigm of AI capability advancement visibly stumbles in its acceleration[3], this type of messaging would become even more effective. The black-box DL paradigm would open itself to derision for being a bubble, an empty promise.

I mention this reluctantly/for comprehensiveness' sake. I think that this is a high-variance approach, most of the attempts at this are going to land badly, and will amount to nothing or have a negative effect. But it is a possible option.

Messaging aimed at the general public is nevertheless a much better, and more neglected, avenue.

  1. ^

    Or maybe not even then, see the Law of Continued Failure.

  2. ^

    The toy model there is roughly:

      Protest 1 is made up of some number of people , who are willing to show their beliefs in public even with the support of zero other people.Protest 2 is joined by  people who are willing to show their beliefs in public if they have the support of  other people....Protest  is joined by  people who are willing to show their beliefs in public if they have the support of  other people.

    (Source, Ctrl+F in the transcript for "second moving part is diverse threshold".)

  3. ^

    Which I do mostly expect. AGI does not seem just around the corner on my inside model of AI capabilities. The current roadmap seems to be "scale inference-time compute, build lots of RL environments, and hope that God will reward those acts of devotion by curing all LLM ailments and blessing them with generalization". Which might happen, DL is weird. But I think there's a lot of room for skepticism with that idea.

    I think the position that The End Is Nigh is being deliberately oversold by powerful actors: the AGI Labs. It's in their corporate interests to signal hype to attract investment, regardless of how well research is actually progressing. So the mere fact that they're acting optimistic carries no information.

    And those of us concerned about relevant X-risks are uniquely vulnerable to buying into that propaganda. Just with the extra step of transmuting the hype into despair. We're almost exactly the people this propaganda is optimized for, after all – and we're not immune to it.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI风险 大众宣传 政治压力 认知提升
相关文章