TechCrunch News 04月16日 03:51
OpenAI says it may ‘adjust’ its safety requirements if a rival lab releases ‘high-risk’ AI
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

OpenAI更新其准备框架,旨在应对日益激烈的AI竞争。新框架允许在竞争对手发布缺乏同等安全措施的高风险系统时,调整其安全要求。此举反映了商业AI开发者加速模型发布的压力。同时,OpenAI强调将更多依赖自动化评估以加快产品开发速度,并更新了模型风险分类标准。这些变化是自2023年以来对准备框架的首次更新,旨在平衡安全性和快速发布的需求。

🛡️ 面对竞争压力,OpenAI可能调整其安全要求。如果其他AI公司发布了没有类似安全措施的高风险系统,OpenAI可能会相应调整其框架。

⚙️ OpenAI正在增加对自动化评估的依赖。为了加快产品开发速度,OpenAI构建了“越来越多的自动化评估套件”,尽管并未完全放弃人工测试,但自动化评估将在更快的模型发布节奏中发挥作用。

⚠️ OpenAI更新了模型风险分类标准。公司将重点关注模型是否达到“高”或“关键”能力阈值,并根据这些风险等级实施相应的安全措施。

In an update to its Preparedness Framework, the internal framework OpenAI uses to decide whether AI models are safe and what safeguards, if any, are needed during development and release, OpenAI said that it may “adjust” its requirements if a rival AI lab releases a “high-risk” system without comparable safeguards.

The change reflects the increasing competitive pressures on commercial AI developers to deploy models quickly. OpenAI has been accused of lowering safety standards in favor of faster releases, and of failing to deliver timely reports detailing its safety testing.

Perhaps anticipating criticism, OpenAI claims that it wouldn’t make these policy adjustments lightly, and that it would keep its safeguards at “a level more protective.”

“If another frontier AI developer releases a high-risk system without comparable safeguards, we may adjust our requirements,” wrote OpenAI in a blog post published Tuesday afternoon. “However, we would first rigorously confirm that the risk landscape has actually changed, publicly acknowledge that we are making an adjustment, assess that the adjustment does not meaningfully increase the overall risk of severe harm, and still keep safeguards at a level more protective.”

The refreshed Preparedness Framework also makes clear that OpenAI is relying more heavily on automated evaluations to speed up product development. The company says that, while it hasn’t abandoned human-led testing altogether, it has built “a growing suite of automated evaluations” that can “keep up with [a] faster [model release] cadence.”

According to the Financial Times, OpenAI gave testers less than a week for safety checks for an upcoming major model — a compressed timeline compared to previous releases. The publication’s sources also alleged that many of OpenAI’s safety tests are now conducted on earlier versions of models than the versions released to the public.

Other changes to OpenAI’s framework pertain to how the company categorizes models according to risk, including models that can conceal their capabilities, evade safeguards, prevent their own shutdown, and even self-replicate. OpenAI says that it’ll now focus on whether models meet one of two thresholds: “high” capability or “critical” capability.

OpenAI’s definition of the former is a model that could “amplify existing pathways to severe harm.” The latter are models that “introduce unprecedented new pathways to severe harm,” per the company.

“Covered systems that reach high capability must have safeguards that sufficiently minimize the associated risk of severe harm before they are deployed,” wrote OpenAI in its blog post. “Systems that reach critical capability also require safeguards that sufficiently minimize associated risks during development.”

The changes are the first OpenAI has made to the Preparedness Framework since 2023.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

OpenAI AI安全 竞争压力 模型发布
相关文章