少点错误 18小时前
AISN #53: An Open Letter Attempts to Block OpenAI Restructuring
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本期AI安全新闻通讯聚焦OpenAI重组争议与CAIS安全基准竞赛。前员工与专家联名致信,呼吁阻止OpenAI重组为营利性机构,认为此举将威胁其慈善使命。信中指出,重组可能削弱AGI的治理保障,使股东利益凌驾于公众利益之上。另一方面,CAIS宣布SafeBench竞赛获奖者,该竞赛旨在评估和降低AI风险,涵盖鲁棒性、监控、对齐和安全应用四大领域,共颁发25万美元奖金,多个基准测试脱颖而出,为AI安全研究提供重要工具。

🛑 OpenAI重组争议:专家和前员工呼吁加州和特拉华州检察长阻止OpenAI重组为营利性机构,认为此举将损害其最初的慈善使命,危及AGI的治理保障,使公司更容易受到利润动机的驱动。

🏆 SafeBench竞赛结果:CAIS的SafeBench竞赛颁发了25万美元奖金,旨在评估和降低AI风险。竞赛关注鲁棒性、监控、对齐和安全应用四大领域,吸引了近百份投稿,获奖基准测试包括Cybench、AgentDojo和BackdoorLLM等。

🛡️ 多个安全基准测试:Cybench评估语言模型在网络安全方面的能力;AgentDojo评估LLM代理的提示注入攻击和防御;BackdoorLLM评估针对大型语言模型的后门攻击;CVE-Bench评估AI代理利用真实网络应用漏洞的能力;JailBreakV评估多模态大型语言模型抵抗越狱攻击的鲁棒性;Poser通过操纵LLM内部结构来检测对齐伪造;Me, Myself, and AI测试LLM的情境感知能力;BioLP-bench评估LLM对生物实验室协议的理解能力。

Published on April 29, 2025 4:13 PM GMT

Welcome to the AI Safety Newsletter by the Center for AI Safety. We discuss developments in AI and AI safety. No technical background required.

In this edition: Experts and ex-employees urge the Attorneys General of California and Delaware to block OpenAI’s for-profit restructure; CAIS announces the winners of its safety benchmarking competition.

Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts.

Subscribe to receive future versions.


An Open Letter Attempts to Block OpenAI Restructuring

A group of former OpenAI employees and independent experts published an open letter urging the Attorneys General (AGs) of California (where OpenAI operates) and Delaware (where OpenAI is incorporated) to block OpenAI’s planned restructuring into a for-profit entity. The letter argues the move would fundamentally undermine the organization's charitable mission by jeopardizing the governance safeguards designed to protect control over AGI from profit motives.

OpenAI was founded with the charitable purpose to ensure that artificial general intelligence benefits all of humanity. OpenAI’s original nonprofit structure, and later its capped-profit model, were designed to control profit motives in the development of AGI, which OpenAI defines as "highly autonomous systems that outperform humans at most economically valuable work." The structure was designed to prevent profit motives from incentivizing OpenAI to take risky development decisions and divert much of the wealth produced by AGI to private shareholders.

The proposed restructuring into a Public Benefit Corporation (PBC) would dismantle the governance safeguards OpenAI originally championed. The letter highlights that the proposed restructuring would transfer control away from the nonprofit entity–whose primary fiduciary duty is to humanity–to a for-profit board whose directors would be partly beholden to shareholder interests. The authors detail several specific safeguards currently in place that would be undermined or eliminated:

The letter concludes by asking the Attorneys General of California and Delaware to halt the restructuring and protect OpenAI’s charitable mission. The authors argue that transferring control of potentially the most powerful technology ever created to a for-profit entity fundamentally contradicts OpenAI's charitable obligations. They urge the AGs to use their authority to investigate the proposed changes and ensure that the governance structures prioritizing public benefit over private gain remain intact.

SafeBench Winners

CAIS recently concluded its SafeBench competition, which awarded prizes for new benchmarks for assessing and reducing risks from AI. Sponsored by Schmidt Sciences, the competition awarded $250,000 across eight winning submissions.

The competition focused on four key areas—Robustness, Monitoring, Alignment, and Safety Applications—attracting nearly one hundred submissions. A panel of judges evaluated submissions based on the clarity of safety assessment, the potential benefit of progress on the benchmark, and the ease of evaluating measurements.

Three Benchmarks Awarded First Prize. Three submissions each received first prizes of $50,000 each for their applicability to frontier models, relevance to current safety challenges, and use of large datasets.

Five Benchmarks Recognized with Second Prize. Five additional submissions were awarded $20,000 each for their innovative approaches to evaluating specific AI safety risks.

These benchmarks provide crucial tools for understanding the progress of AI, evaluating risks, and ultimately reducing potential harms. The papers, code, and datasets for all winning benchmarks are publicly available for further research and use. CAIS hopes to see future work which is inspired by or builds on these submissions.

Other News

Government

Research and Opinion

AI Frontiers


See also: CAIS website, X account for CAIS, our paper on superintelligence strategy, our AI safety course, and AI Frontiers, a new platform for expert commentary and analysis.

Listen to the AI Safety Newsletter for free on Spotify or Apple Podcasts.

Subscribe to receive future versions.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI安全 OpenAI 安全基准 AGI治理
相关文章