少点错误 16小时前
RTFB: The RAISE Act
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文深入探讨了纽约州RAISE法案,该法案旨在规范人工智能(AI)领域,特别是针对前沿模型开发商。文章详细解读了法案的核心内容,包括对“前沿模型”的定义、安全协议的制定与公开、以及安全事件的报告机制。通过分析法案的关键条款和潜在影响,文章旨在帮助读者全面了解这项立法,并评估其对AI行业及更广泛社会的影响。文章还提到了该法案在纽约州议会和参议院中获得压倒性通过,以及州长Kathu Hochul是否签署该法案的决定。

🤖**前沿模型定义:** RAISE法案将“前沿模型”定义为,训练成本超过1亿美元,或经过知识蒸馏后成本超过500万美元的AI模型。该定义旨在针对少数大型AI开发者,而非影响广泛的初创企业。

🚨**安全协议要求:** 法案要求前沿模型开发者制定、公开并遵守安全和安保协议(SSP),以降低模型可能造成的严重危害风险。这包括对可能造成100人伤亡或10亿美元损失的事件进行预防。

📢**安全事件报告:** 法案规定,在发生安全事件后的72小时内,开发者必须向纽约州总检察长和国土安全与应急服务部报告。报告内容包括事件日期、性质及简要描述,类似于网络安全事件的报告标准。

⚖️**执法与处罚:** 法案的执法由总检察长负责,违规行为可能导致禁令救济和罚款。首次违规罚款最高1000万美元,后续违规最高3000万美元,旨在促使开发者积极采取预防措施。

Published on June 16, 2025 12:50 PM GMT

The RAISE Act has overwhelmingly passed the New York Assembly (95-1 among Democrats and 24-21 among Republicans) and New York Senate (37-1 among Democrats, 21-0 among Republicans).Governor Kathy Hochul now has to decide whether or not to sign it, which she has 10 non-Sunday days to do once the bill is delivered (30 if they’re out of session), but the bill might not be delivered for six months.The aim of this post, now that we are seeing increasing public discussion, is to go through the bill to understand exactly what the bill would and would not do.

Overall Take

The RAISE Act is centrally a transparency bill. It requires frontier model developers to maintain, publish and adhere to (one might say ‘open source’ except that they can redact details for various reasons) a safety and security protocol (SSP) that outlines how they will, before releasing their frontier models, take appropriate steps to reduce risk of critical harm (100 casualties or 1 billion in damages) caused or materially enabled by those models. It must designate senior people as responsible for implementation.
It also requires companies to disclose (as in, write two sentences informing us about) safety incidents within 72 hours.Enforcement is done only by the attorney general, and limited to injunctive or declaratory relief and fines of a maximum of $10 million for the first violation and $30 million for subsequent violations. This can happen if a company fails to take appropriate preventative steps, even if no critical harm has yet resulted, so if the SSP proves sufficiently inadequate preemptive action can be taken.My take on the RAISE Act is that it seems clearly to be bending over backwards to avoid imposing substantial costs on the companies involved even if the state were to attempt to enforce it maximally and perversely, to give those companies maximum flexibility in how they respond, and to only apply to a handful of major players.The bill is thus insufficient on its own but an important improvement upon the status quo. I strongly support this bill. I am very much not alone. The RAISE Act is a highly popular bill, supported (with admittedly very low salience) by 84% of New Yorkers.a16z has already attempted to kill this bill before it overwhelmingly passed both houses, circulating an opposition memo and reportedly calling members. We should expect a continued flurry of industry lobbying against RAISE, likely following the usual playbooks, and for them to greatly outspend bill advocates.o3-pro thinks Hochul is likely to ultimately sign the bill. with a 65% chance it becomes law in current form, 15% chance it becomes law with negotiated chapter amendments. The Manifold market has a 57% chance that the bill becomes law.

RTFB: The RAISE Act

There are two big advantages we have in reading the RAISE Act.
    It is short and simple. We’ve analyzed similar things before.
Relax. This will be a breeze.The bill is mostly definitions.

Definitions: Frontier Model

These are mostly standard. The AI definition has been consistent for a while. Compute cost is defined as the published market price cost of cloud compute, as reasonably assessed by the person doing the training, which is as clear and generous as one could hope.The most important definition is ‘frontier model’:
6. “Frontier model” means either of the following:(a) an artificial intelligence model trained using greater than 10^26 computational operations (e.g., integer or floating-point operations), the compute cost of which exceeds one hundred million dollars;OR(b) an artificial intelligence model produced by applying knowledge distillation to a frontier model as defined in paragraph (a) of this subdivision, provided that the compute cost for such model produced by applying knowledge distillation exceeds five million dollars.
The first provision will centrally be ‘you spent $100 million dollars.’ Which remains a lot of dollars, and means this will only apply to a handful of frontier labs. But also note that 10^26 will for a while remain a lot of FLOPS. Epoch looked at this question, and also estimates the costs of various models, with the only current model over 10^26 likely being Grok 3 (o3-pro suggests it is not impossible that Gemini Ultra or a few others might just barely also qualify, although I find this highly unlikely).The question is the second provision. How often will companies make distillations that cost more than $5 million and result in ‘similar or equivalent capabilities’ to the original, as required by the definition of distillation used here?o3-pro believes the current number of such models, even without considering the capabilities requirement, is probably zero (the possible exception is Claude Haiku, if you think it has sufficiently comparable capabilities). It anticipates the number of $5 million distillations will not remain zero, and expects the distillations to mostly (but not entirely) be from the same companies releasing the $100 million frontier models.Its baseline scenario is by 2029, there will be ~6 American frontier-trainers, in particular OpenAI, DeepMind, Anthropic, Meta, xAI and then maybe Amazon or Apple or perhaps an open source collective, and ~6 more distillers on top of that passing the $5 mark, starting with Cohere, then maybe Databricks or Perplexity.A ‘large developer’ means spending a combined $100 million in training compute, or someone who buys the full intellectual rights to the results of that, with academic institutions doing research excluded.This bill would have zero impact on everyone else.So yes, there will be talk about how this will be ‘more difficult’ for ‘smaller’ companies. But by ‘smaller’ companies we mean a handful of large companies, and by ‘more difficult’ we mean a tiny fraction of overall costs. And as always, please point to the thing that you would have to do, that you don’t think is worth doing, or is even a substantial impact on their business costs?Bill opponents, of course, are telling the same lies about this they told about SB 1047. Brianna January of the ‘Chamber of Progress’ calls this ‘an eviction notice for New York’s 9,000 AI startups,’ saying it ‘would send AI innovators packing,’ when exactly zero of these 9,000 startups would have to lift a single finger in response to this bill.This is pure bad faith Obvious Nonsense, and you should treat anyone who says similar things accordingly. (The other Obvious Nonsense claim here is that the bill was ‘rushed’ and lacked a public hearing. The bill very much followed normal procedures and had debate on the floor, the bill was in the public pipeline for months, and bills in New York do not otherwise get pubic hearings, that’s a non sequitur.)

Definitions: Critical Harm

“Critical harm” means the death or serious injury of one hundred or more people or at least one billion dollars of damages to rights in money or property caused or materially enabled by a large developer’s use, storage, or release of a frontier model, through either of the following:(a) The creation or use of a chemical, biological, radiological, or nuclear weapon; or(b) An artificial intelligence model engaging in conduct that does both of the following:(i) Acts with no meaningful human intervention; and(ii) Would, if committed by a human, constitute a crime specified in the penal law that requires intent, recklessness, or gross negligence, or the solicitation or aiding and abetting of such a crime.A harm inflicted by an intervening human actor shall not be deemed to result from a developer’s activities unless such activities were a substantial factor in bringing about the harm, the intervening human actor’s conduct was reasonably foreseeable as a probable consequence of the developer’s activities, and could have been reasonably prevented or mitigated through alternative design, or security measures, or safety protocols.
We have ‘caused or materially enabled’ and also ‘substantial factor’ and ‘harm that mitigations could have reasonably prevented’ and either 100 serious injuries or a billion dollars in damage as the thresholds, and either the act has to be autonomous, be a CBRN risk, or constitute a crime in the penal law.That seems like a robust way of saying ‘if you trigger this provision you screwed up?’

Definition: Safety Incident

They have to be reported, so what exactly are they?
“Safety incident” means a known incidence of critical harmOR an incident of the following kinds that occurs in such a way that it provides demonstrable evidence of an increased risk of critical harm:
    A frontier model autonomously engaging in behavior other than at the request of a user; Theft, misappropriation, malicious use, inadvertent release, unauthorized access, or escape of the model weights of a frontier model; The critical failure of any technical or administrative controls, including controls limiting the ability to modify a frontier model; Unauthorized use of a frontier model.
The incidence of an actual critical harm is clear.The second half of the definition has two halves.
    It has to involve one of the four things listed. It has to provide demonstrable evidence of an increased risk of critical harm.
As in, something in your safety protocols goes wrong, in a way that makes you more worried about risk. That seems like the kind of thing you should report. I will be very happy to see these systematically written down, and even happier to have them disclosed.

You Have To Report Safety Incidents The Way We Report Cybersecurity Breaches

As in, within 72 hours of any safety incidents, you have to notify the attorney general and DHSES. This is the common standard used for cybersecurity breaches. You have to include:
    The date of the incident. Why it qualifies as a safety incident. ‘A short and plain statement describing the safety incident.’
Does this, as some have suggested, constitute such a burden that it interferes with the ability to respond to the incident? That seems difficult to believe.For example, you could write ‘On Tuesday, June 17, 2025, someone gained unauthorized access to our frontier model. This makes us more worried about future unauthorized access.’ That’s it.I have no sympathy for the claim that asking for that style of statement within three days is a distracting or undue burden that outweighs our right to know, or its costs exceed benefits. In many cases, waiting longer could have serious repercussions.

Definition: Safety and Security Protocol (SSP)

What are we actually asking companies to produce, exactly? A documentation and description of technical and organizational protocols that if fully implemented would:
    ‘Appropriately reduce the risk of critical harm.’ ‘Appropriately reduce the risk of’ unauthorized access to or misuse of the model weights ‘leading to critical harm.’ Describe a detailed test procedure to evaluate potential misuse or loss of control or combination with other software to potentially cause critical harm. Enable compliance with this article. Designate senior personnel to be responsible for ensuring compliance.
This requires ‘detailed test procedures’ to be described in advance, which seems like a very good idea, and does not preclude additional tests. The rest seems so basic that it seems laughable to object to being told to do any of it.

You Have To Have And Implement An SSP

Before deploying (meaning externally, as in giving a third party access) to a frontier model, the developer must write and implement an SSP, retain an up-to-date copy of that SSP, conspicuously publish a redacted copy of the SSP, give the attorney general and DHSES access upon request to the full SSP and retain copies of all your test results sufficient to allow third-party replication.As always, there are no specific requirements for the SSP, other than that it must ‘appropriately reduce the risk’ of critical harms, both directly or through unauthorized access, and that it spell out your testing procedure, and that you actually have someone ensure you use it. If you want to write the classic ‘lol we’re Meta, we don’t run tests, full open weights release without them seems appropriate, I’m sure it will be fine’ you can do that, although you might not like what happens when people notice you did that, or when the risks materialize, or potentially the AG notices you’re not taking the appropriate actions and sues you.You need to conduct an annual review of the SSP to adjust for increased model capabilities, and make and publish any appropriate adjustments. Seems wise.

You Have To Not Release Models That Would Create An Unreasonable Risk Of Critical Harm

I for one think that if your model would create an unreasonable risk of critical harm then that means you shouldn’t release it. But that’s just me.

You Shall Not Knowingly Make False Or Materially Misleading Statements or Omissions

Again, yeah, I mean, I hope that stands to reason.

What Are The Consequences?

The attorney general can bring a civil action with penalties of:
    $10 million for the first violation, $30 million for additional ones. Injunctive or declaratory relief.
And that’s it. Explicitly no private right of action, no limit of the application of other laws, everything is cumulative with other requirements. If you cause an incident that costs billions of dollars, your fines don’t scale with that.I don’t see any clause allowing compensatory relief. So if there’s a violation related to an actual critical harm, I presume any fines involved will be the least of your problems.The main actual consequences are that frontier labs will be forced to be transparent about their safety and security protocols (SSPs) and what tests they intend to run and other precautions they intend to take, in order to guard against critical harms. Most labs impacted already do this, and will only have to newly include the evals they intend to run. Publishing these details will allow us to critique them, and apply pressure to create better protocols.Again, while I have concerns that the bill is insufficient strong, I think all of this is a very good thing. I strongly support the bill.  

Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

RAISE法案 人工智能 监管 安全协议 纽约州
相关文章