少点错误 2024年11月05日
Anthropic - The case for targeted regulation
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

随着AI系统能力的快速发展,其潜在风险也日益凸显,包括网络安全和生物化学领域。文章指出,AI在编码、网络攻击和科学理解方面取得了显著进步,这同时也带来了潜在的破坏性应用风险。作者呼吁政府在未来18个月内采取行动,制定明智的AI政策,以平衡AI带来的益处和风险,避免因监管滞后导致的负面后果。文章还强调了AI在网络安全和CBRN领域带来的潜在风险,并建议政府采取措施,在支持AI创新发展的同时,有效降低其带来的灾难性风险。

🤔 **AI能力快速提升带来风险**: AI系统在数学、推理、编码等方面取得飞速进步,但也可能被用于网络安全攻击或生物化学领域,带来潜在破坏。

💻 **网络安全风险加剧**: AI模型在软件工程和网络攻击评估方面的能力显著提升,未来能够规划复杂多步骤任务的AI模型可能会更有效地用于网络攻击。

🦠 **CBRN领域潜在风险**: AI模型能够获取生物化学领域的专业知识,其在科学理解方面的进步也可能被用于CBRN领域的恶意用途,风险正在快速增长。

⚠️ **风险临近,需紧急采取行动**: 作者认为,AI在网络安全和CBRN领域带来的风险正在快速逼近,需要政府在未来18个月内采取措施,制定明智的AI政策。

💡 **建议:谨慎且有针对性的监管**: 文章建议政府采取谨慎且有针对性的监管措施,以平衡AI带来的益处和风险,避免因监管滞后而导致的负面后果。

Published on November 5, 2024 7:07 AM GMT

The first two sections are below:

Increasingly powerful AI systems have the potential to accelerate scientific progress, unlock new medical treatments, and grow the economy. But along with the remarkable new capabilities of these AIs come significant risks. Governments should urgently take action on AI policy in the next eighteen months. The window for proactive risk prevention is closing fast.

Judicious, narrowly-targeted regulation can allow us to get the best of both worlds: realizing the benefits of AI while mitigating the risks. Dragging our feet might lead to the worst of both worlds: poorly-designed, knee-jerk regulation that hampers progress while also failing to be effective at preventing risks.

In this post, we suggest some principles for how governments can meaningfully reduce catastrophic risks while supporting innovation in AI’s thriving scientific and commercial sectors.

Urgency

In the last year, AI systems have grown dramatically better at math, graduate-level reasoning, and computer coding, along with many other capabilities. Inside AI companies, we see continued progress on as-yet undisclosed systems and results. These advances offer many positive applications. But progress in these same broad capabilities also brings with it the potential for destructive applications, either from the misuse of AI in domains such as cybersecurity or biology, or from the accidental or autonomous behavior of the AI system itself.

In the realm of cyber capabilities, models have rapidly advanced on a broad range of coding tasks and cyber offense evaluations. On the SWE-bench software engineering task, models have improved from being able to solve 1.96% of a test set of real-world coding problems (Claude 2, October 2023) to 13.5% (Devin, March 2024) to 49% (Claude 3.5 Sonnet, October 2024). Internally, our Frontier Red Team has found that current models can already assist on a broad range of cyber offense-related tasks, and we expect that the next generation of models—which will be able to plan over long, multi-step tasks—will be even more effective.

On the potential for AI exacerbating CBRN (chemical, biological, radiological, and nuclear) misuses, the UK AI Safety Institute tested a range of models from industry actors (including Anthropic) and concluded that:

...models can be used to obtain expert-level knowledge about biology and chemistry. For several models, replies to science questions were on par with those given by PhD-level experts.

AI systems have progressed dramatically in their understanding of the sciences in the last year. The widely used benchmark GPQA saw scores on its hardest section grow from 38.8% when it was released in November 2023, to 59.4% in June 2024 (Claude 3.5 Sonnet), to 77.3% in September (OpenAI o1; human experts score 81.2%). Our Frontier Red Team has also found continued progress in CBRN capabilities. For now, the uplift of having access to a frontier model relative to existing software and internet tools is still relatively small, however it is growing rapidly. As models advance in capabilities, the potential for misuse is likely to continue on a similar scaling trend.

About a year ago, we warned that frontier models might pose real risks in the cyber and CBRN domains within 2-3 years. Based on the progress described above, we believe we are now substantially closer to such risks. Surgical, careful regulation will soon be needed.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

人工智能 AI风险 AI政策 网络安全 CBRN
相关文章