少点错误 02月13日
Skepticism towards claims about the views of powerful institutions
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

当前AI治理领域面临诸多挑战,各国政府更迭、技术发展迅速以及政治立场的转变,使得解读相关政策和行动变得复杂。本文分析了在评估机构对AI的看法时,可能出现的认知偏差。例如,当有人声称某个机构的观点与自己一致时,我们应保持警惕,因为这可能受到个人偏见、战略意图或利益驱动的影响。同样,对声称机构与自己观点相悖的说法,也应审慎对待。因此,在AI治理的实践中,我们需要结合自身判断和多方证据,避免盲从他人观点。

🤝 **人际关系偏差:** 人们与机构内部人员的互动可能存在偏差。因为Alice认识的机构内部人员可能与她有相似的背景或观点,导致她高估了机构对她的支持度。

🎭 **战略意图:** Alice可能出于战略目的而声称机构与她观点一致,以增加自身的影响力或促使机构采纳她的建议。这种行为带有一定的目的性,可能扭曲事实。

🔮 **确认偏差:** 如果Alice曾预测该机构会支持她的观点,她可能会受到确认偏差的影响,倾向于寻找和解释支持她预测的信息,而忽略或淡化相反的证据。

🎯 **影响对立机构:** Bob可能会声称某个机构(A)与他的观点不一致,以此来展示他对该机构的反对,并加强与对立机构(B)的联系,从而影响B机构的立场。

💰 **利益驱动:** Alice的工作可能旨在影响某个机构,为了更容易地招募人员和筹集资金,她可能会夸大自己对该机构的影响力,从而使人们相信影响该机构是可行的。

Published on February 13, 2025 7:40 AM GMT

Introduction: some contemporary AI governance context

It’s a confusing time in AI governance. Several countries’ governments recently changed hands. DeepSeek and other technical developments have called into question certain assumptions about the strategic landscape. Political discourse has swung dramatically away from catastrophic risk and toward framings of innovation and national competitiveness.

Meanwhile, the new governments have issued statements of policy, and AI companies (mostly) continue to publish or update their risk evaluation and mitigation approaches. Interpreting these words and actions has become an important art for AI governance practitioners: does the phrase “human flourishing” in the new executive order signal concern about superintelligence, or just that we should focus on AI’s economic and medical potential and not “hand-wring” about safety? How seriously should we take the many references to safety in the UK’s AI Opportunities Action Plan, given the unreserved AI optimism in the announcement? Does Meta’s emphasis on “unique” risks take into account whether a model’s weights are openly released? The answers matter not only for predicting future actions but also for influencing them: it’s useful to know an institution’s relative appetite for different kinds of suggestions, e.g. more export controls versus maintaining Commerce’s reporting requirements.

So, many people who work in AI governance spend a lot of time trying to read between the lines of these public statements, talking to their contacts at these institutions, and comparing their assessment of the evidence with others’. This means they can wind up with a lot of non-public information — and often, they also have lots of context that casual observers (or people who are doing heads-down technical work in the Bay) might not.

All of that is to say: if you hear someone express a view about how an institution is thinking about AI (or many other topics), you might be tempted to update your own view towards theirs, especially if they have expertise or non-public information. And, of course, this is sometimes the correct response.

But this post argues that you should take these claims with a grain of salt. The rest of the post shifts to a much higher level of abstraction than the above, in part because I don’t want to “put anyone on blast,” and in part because this is a general phenomenon. Note that lots of these are generic reasons to doubt claims you can’t independently verify, but some of them are specific to powerful institutions.

Biases towards claiming agreement with one’s own beliefs

Let’s say you hear Alice say that a powerful institution (like a political party, important company, government, etc.) agrees with her position on a controversial topic more than you might think.

If you have reason to think that Alice knows more about that institution than you do, or just has some information that you don’t have, you might be inclined to believe Alice and update your views accordingly: maybe that institution is actually more sympathetic to Alice’s views than you realized!

This might be true, of course. But I’d like to point out a few reasons to be skeptical of this claim.

Weaker biases towards claiming disagreement with one’s own beliefs

Now imagine that you hear Bob, who agrees with Alice’s view, make the opposite claim: actually, the institution disagrees with us!

Not all of the same factors above apply – and I think, on net, these effects are stronger for those claiming agreement than disagreement, roughly in proportion to how powerful the institution is. But some of them still do, at least for some permutation:

Conclusion

I wouldn’t totally dismiss either claim, especially if Alice/Bob do have some private information, even if I knew that they had many of these biases. Claims like theirs are a valuable source of evidence. But I would take both claims (especially Alice’s) with a grain of salt, and if the strength of these claims were relevant for an important decision, I’d consider whether and to what extent these biases might be at play. This means giving a bit more weight to my own prior views of the institution and my own interpretations of the evidence, albeit only to the extent that I think biases like the above apply less to me than to the source of the claim.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI治理 认知偏差 政策解读 战略分析 信息评估
相关文章