TechCrunch News 前天 02:56
Google’s latest AI model report lacks key safety details, experts say
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

谷歌发布了其最强大的AI模型Gemini 2.5 Pro的内部安全评估报告,但专家指出,该报告信息量有限,难以评估模型潜在风险。尽管谷歌承诺公开AI安全报告,但此次报告的细节匮乏,引发了对谷歌履行其公开承诺的质疑。与其他AI巨头如Meta和OpenAI类似,谷歌在安全报告的发布上似乎有所保留。专家们呼吁更及时、更全面的安全评估,以确保AI模型的安全性和透明度,避免在AI安全方面出现“恶性竞争”。

🧐 谷歌发布了关于其最强大的AI模型Gemini 2.5 Pro的内部安全评估报告,但专家认为报告细节不足,难以评估模型潜在风险。

🤔 谷歌的安全报告策略与竞争对手有所不同,只有当模型“走出实验阶段”后才会发布技术报告,并且不会在报告中包含所有“危险能力”评估结果。

😮 专家们对报告的稀疏性表示失望,并指出报告未提及谷歌的Frontier Safety Framework (FSF),该框架旨在识别可能造成“严重损害”的未来AI能力。

😟 谷歌在AI安全报告方面的表现未达到承诺,未能及时提供补充安全评估,且未发布Gemini 2.5 Flash模型的报告。

😨 专家认为,这种信息披露的不足,以及其他AI实验室在发布前缩短安全测试时间的做法,可能导致AI安全和透明度的“恶性竞争”。

On Thursday, weeks after launching its most powerful AI model yet, Gemini 2.5 Pro, Google published a technical report showing the results of its internal safety evaluations. However, the report is light on the details, experts say, making it difficult to determine which risks the model might pose.

Technical reports provide useful — and unflattering, at times — info that companies don’t always widely advertise about their AI. By and large, the AI community sees these reports as good-faith efforts to support independent research and safety evaluations.

Google takes a different safety reporting approach than some of its AI rivals, publishing technical reports only once it considers a model to have graduated from the “experimental” stage. The company also doesn’t include findings from all of its “dangerous capability” evaluations in these write-ups; it reserves those for a separate audit.

Several experts TechCrunch spoke with were still disappointed by the sparsity of the Gemini 2.5 Pro report, however, which they noted doesn’t mention Google’s Frontier Safety Framework (FSF). Google introduced the FSF last year in what it described as an effort to identify future AI capabilities that could cause “severe harm.”

“This [report] is very sparse, contains minimal information, and came out weeks after the model was already made available to the public,” Peter Wildeford, co-founder of the Institute for AI Policy and Strategy, told TechCrunch. “It’s impossible to verify if Google is living up to its public commitments and thus impossible to assess the safety and security of their models.”

Thomas Woodside, co-founder of the Secure AI Project, said that while he’s glad Google released a report for Gemini 2.5 Pro, he’s not convinced of the company’s commitment to delivering timely supplemental safety evaluations. Woodside pointed out that the last time Google published the results of dangerous capability tests was in June 2024 — for a model announced in February that same year.

Not inspiring much confidence, Google hasn’t made available a report for Gemini 2.5 Flash, a smaller, more efficient model the company announced last week. A spokesperson told TechCrunch a report for Flash is “coming soon.”

“I hope this is a promise from Google to start publishing more frequent updates,” Woodside told TechCrunch. “Those updates should include the results of evaluations for models that haven’t been publicly deployed yet, since those models could also pose serious risks.”

Google may have been one of the first AI labs to propose standardized reports for models, but it’s not the only one that’s been accused of underdelivering on transparency lately. Meta released a similarly skimpy safety evaluation of its new Llama 4 open models, and OpenAI opted not to publish any report for its GPT-4.1 series.

Hanging over Google’s head are assurances the tech giant made to regulators to maintain a high standard of AI safety testing and reporting. Two years ago, Google told the U.S. government it would publish safety reports for all “significant” public AI models “within scope.” The company followed up that promise with similar commitments to other countries, pledging to “provide public transparency” around AI products.

Kevin Bankston, a senior adviser on AI governance at the Center for Democracy and Technology, called the trend of sporadic and vague reports a “race to the bottom” on AI safety.

“Combined with reports that competing labs like OpenAI have shaved their safety testing time before release from months to days, this meager documentation for Google’s top AI model tells a troubling story of a race to the bottom on AI safety and transparency as companies rush their models to market,” he told TechCrunch.

Google has said in statements that, while not detailed in its technical reports, it conducts safety testing and “adversarial red teaming” for models ahead of release.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

谷歌 AI安全 透明度 Gemini 2.5 Pro 安全报告
相关文章