少点错误 06月05日 18:52
Powerful Predictions
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了人工智能(AI)领域中预测在政策制定中的应用,以及如何通过预测来影响AI的未来发展。文章调查了多个致力于AI安全和政策的组织,分析了它们如何将预测融入决策、政策建议和战略规划中。研究发现,虽然预测在AI政策制定中可能被低估,但一些组织,如Open Philanthropy和Anthropic,已经开始积极运用预测来评估项目结果和AI风险。此外,文章还提到了多个进行AI预测的组织,例如MIRI、Epoch AI等,它们通过不同的方式影响着政策制定和公众认知,为AI的未来发展奠定了基础。

💡Open Philanthropy会定期对资助项目的成果进行预测,以此来提升未来资助决策的准确性。他们还委托Arb Research进行AI预测,尽管具体应用未公开。

💡Anthropic在其负责任的规模化政策(RSP)中,通过预测来全面评估AI,尤其关注诱导技术和模型性能的提升。他们致力于不断改进预测,以便在风险评估中提供更可靠的依据。

💡MIRI在2023年调整了策略,转向AI治理和沟通,并制定了涉及四个高级别情景的战略框架,以应对重要的治理研究问题。他们希望通过各种渠道影响政策制定者、顾问和公众。

💡Epoch AI作为AI预测组织,研究AI的发展轨迹及其对经济和社会的影响。他们与政策制定者合作,为英国政府和欧盟委员会提供咨询和合作,并参与了英国和荷兰的AI相关政策制定。

💡其他组织,如IAPS、RAND、CSET和TFI,也通过预测和研究影响AI政策。他们通过举办研讨会、发布研究报告、参与国会听证等方式,将研究成果传递给政策制定者和公众。

Published on June 5, 2025 10:44 AM GMT

A thoughtful post by Anton Leicht, “Powerless Predictions,” notes that forecasting may be underutilized by AI policy organizations. But how much is forecasting actually being used? Which organizations working on AI—particularly AI safety—already integrate forecasting into their decision-making, policy recommendations, and strategic planning? What forecasting work is being done, and how is it impacting the future of AI?

I set out to investigate this.

While I hope this post proves useful to others, I wrote it primarily for myself to gain an overview of forecasting work and its impacts. Feel free to skim—only read the details if you find them valuable.

Grantmaking

Open Philanthropy, which funds numerous AI safety initiatives, regularly makes predictions about grant outcomes. Evaluating these outcomes helps them improve their ability to predict the results of future grantmaking decisions. Javier Prieto at Open Philanthropy explains this process and evaluates prediction accuracy in this post.

They have also commissioned an long list of forecasting questions on AI from Arb Research. I couldn't find any publicly documented uses of this list, though Open Philanthropy may have used it for internal planning and strategic thinking.

AI development

Anthropic extrapolates rare AI behaviors to reduce concerns about concerning behaviors that may be missed during evaluations.

In Anthropic's Responsible Scaling Policy (RSP), they incorporate forecasting for comprehensive AI assessment, making informal predictions about improvements of elicitation techniques and enhanced model performance between testing rounds. They aim to "improve these forecasts over time so that they can be relied upon for risk judgments."

While you could argue that frontier AI development would be irresponsible without forecasting future capabilities and risks, it's nevertheless encouraging to see this explicitly incorporated into their practices. DeepMind's Frontier Safety Framework doesn't appear to explicitly include forecasting, and OpenAI's Preparedness Framework only mentions it in passing.

Affecting Policy Decisions

While policy recommendations implicitly depend on predictions about the future, explicit and well-researched forecasting doesn't appear to be the norm. However, much policy work may be informed by forecasts without being transparently based on forecast analyses—making it difficult to determine how extensively forecasts are used in practice.

It's also not always clear how organizations actually engage with policymakers. Some report specific engagements, like policy recommendations sent to particular institutions or congressional testimonies, while others simply report general strategies.

I compiled a list of organizations doing forecasting while trying to influence AI policy in a positive direction[1]:

Public awareness and opening the Overton Window

The AI 2027 scenario leveraged multiple channels to reach a broad audience—including promotion through Astral Codex Ten, podcasts, social media, and a website with compelling visual design.

Leopold Aschenbrenner's Situational Awareness similarly demonstrated forecasting's potential for influencing public discourse through accessible language while investigating the future of AI.

Some policy organizations mentioned earlier aim to make their research accessible to the general public, often as an indirect approach to influencing policy—broadening the Overton Window and relying public pressure policymakers into encourage sensible policymaker decisions.

There’s also a collection of blogs investigating the future of AI and society, including the AI Futures Project blog, Sentinel Global Risks Watch, Foxy Scout, and my own blog Forecasting AI Futures.

Foundational Forecasting Research

Some AI forecasting work is key to building a foundation for further research and applied efforts like positive policy influence.

Notable examples include:

You could include AI benchmark work as foundational forecasting research even though it's not directly about forecasting, since benchmark performance trends are very helpful for predicting future capabilities.

I worry that foundational work may struggle to reach key decision-makers who could most benefit from the insights.

With Claude Sonnet 4’s help I found three explicit references to Epoch AI research in policy documents (mentioned earlier),which shows their work is reaching some policymakers even outside direct engagements—though this is still very limited. Admittedly, some policy work may be implicitly informed by forecasting research without explicit references.

Final words

I believe my own work through the Forecasting AI Futures blog falls mostly in the foundational category—the focus has been on gaining a better understanding of key dynamics and potential outcomes. While I want to make it accessible to a broader audience and share it with key people, these objectives haven't been the primary focus.

I hope to work toward forecasting with more direct applications. It seems too easy to fall into the trap of investigating things that seem somewhat important while hoping someone will notice and use the insights.

I still feel uncertain in what these results say about how I should proceed—but at least I have a better overview and foundation for strategizing further. I should probably reach out to various organizations and people more.

I still feel uncertain about what this investigation into forecasting says regarding how I should proceed—but at least I have a better overview and foundation for further strategizing.


Thank you for reading! If you found value in this post, consider subscribing!

  1. ^

    I would have liked to investigate examples where forecasting analyses have been explicitly referenced and used in recommendations and in policymaker engagement, but this would require much more time than I’m willing to spend. Instead, I investigated which organizations that are doing both forecasting and influencing policy (with the exception of MIRI, which explicitly uses forecasting for prioritizing their efforts.)



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI预测 AI政策 机器学习
相关文章