TechCrunch News 03月13日 00:11
Sakana claims its AI-generated paper passed peer review — but it’s a bit more nuanced than that
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

日本初创公司Sakana声称其AI生成了首篇经过同行评审的科学出版物,但这一说法存在一些重要的注意事项。尽管AI在撰写听起来像人类的文章方面表现出色,但其技术缺陷,如产生幻觉的倾向,使许多科学家对其在严肃工作中的应用持谨慎态度。Sakana公司坦承其AI有时会犯令人尴尬的引用错误,且论文并未像其他同行评审出版物那样受到严格审查。有专家指出,AI擅长向人类推销想法,通过同行评审并不等同于为该领域贡献知识。Sakana公司表示,该实验旨在研究AI生成研究的质量,并强调迫切需要制定关于AI生成科学的规范。

🤖 Sakana公司的AI系统The AI Scientist-v2生成了一篇论文,并提交给ICLR的研讨会,该研讨会同意合作进行AI生成稿件的双盲评审实验。

🧪 该AI系统能够“端到端”地生成论文,包括科学假设、实验和实验代码、数据分析、可视化、文本和标题。其中一篇论文被ICLR研讨会接受,但Sakana公司为了透明度和尊重ICLR惯例,在发表前撤回了该论文。

⚠️ 尽管论文被接受,但Sakana承认其AI偶尔会犯“令人尴尬”的引用错误,且论文并未像其他同行评审出版物那样受到严格审查,因为该公司在初步同行评审后撤回了该论文,没有接受额外的“元评审”。

🤔 专家们对AI在科学研究中的作用持不同看法。一些人认为AI尚未准备好担任“合作科学家”,另一些人则认为它具有潜力,但承认目前尚处于早期阶段。此外,专家们担心AI可能会在科学文献中产生噪音,而不是促进进步。

Japanese startup Sakana said that its AI generated the first peer-reviewed scientific publication. But while the claim isn’t untrue, there are significant caveats to note.

The debate swirling around AI and its role in the scientific process grows fiercer by the day. Many researchers don’t believe AI is quite ready to serve as a “co-scientist,” while others think that there’s potential — but acknowledge it’s early days.

Sakana falls into the latter camp.

The company said that it used an AI system called The AI Scientist-v2 to generate a paper that Sakana then submitted to a workshop at ICLR, a long-running and reputable AI conference. Sakana claims that the workshop’s organizers, as well as ICLR’s leadership, had agreed to work with the company to conduct an experiment to double-blind review AI-generated manuscripts.

Sakana said it collaborated with researchers at the University of British Columbia and the University of Oxford to submit three AI-generated papers to the aforementioned workshop for peer review. The AI Scientist-v2 generated the papers “end-to-end,” Sakana claims, including the scientific hypotheses, experiments and experimental code, data analyses, visualizations, text, and titles.

“We generated research ideas by providing the workshop abstract and description to the AI,” Robert Lange, a research scientist and founding member at Sakana, told TechCrunch via email. “This ensured that the generated papers were on topic and suitable submissions.”

One paper out of the three was accepted to the ICLR workshop — a paper that casts a critical lens on training techniques for AI models. Sakana said it immediately withdrew the paper before it could be published in the interest of transparency and respect for ICLR conventions.

A snippet of Sakana’s AI-generated paper.Image Credits:Sakana

“The accepted paper both introduces a new, promising method for training neural networks and shows that there are remaining empirical challenges,” Lange said. “It provides an interesting data point to spark further scientific investigation.”

But the achievement isn’t as impressive as it might seem at first glance.

In a blog post, Sakana admits that its AI occasionally made “embarrassing” citation errors, for example incorrectly attributing a method to a 2016 paper instead of the original 1997 work.

Sakana’s paper also didn’t undergo as much scrutiny as some other peer-reviewed publications. Because the company withdrew it after the initial peer review, the paper didn’t receive an additional “meta-review,” during which the workshop organizers could have in theory rejected it.

Then there’s the fact that acceptance rates for conference workshops tend to be higher than acceptance rates for the main “conference track” — a fact Sakana candidly mentions in its blog post. The company said that none of its AI-generated studies passed its internal bar for ICLR conference track publication.

Matthew Guzdial, an AI researcher and assistant professor at the University of Alberta, called Sakana’s results “a bit misleading.”

“The Sakana folks selected the papers from some number of generated ones, meaning they were using human judgment in terms of picking outputs they thought might get in,” he said via email. “What I think this shows is that humans plus AI can be effective, not that AI alone can create scientific progress.”

Mike Cook, a research fellow at King’s College London specializing in AI, questioned the rigor of the peer reviewers and workshop.

“New workshops, like this one, are often reviewed by more junior researchers,” he told TechCrunch. “It’s also worth noting that this workshop is about negative results and difficulties — which is great, I’ve run a similar workshop before — but it’s arguably easier to get an AI to write about a failure convincingly.”

Cook added that he wasn’t surprised an AI can pass peer review, considering that AI excels at writing human-sounding prose. Partly-AI-generated papers passing journal review isn’t even new, Cook pointed out, nor are the ethical dilemmas this poses for the sciences.

AI’s technical shortcomings — such as its tendency to hallucinate — make many scientists wary of endorsing it for serious work. Moreover, experts fear AI could simply end up generating noise in the scientific literature, not elevating progress.

“We need to ask ourselves whether [Sakana’s] result is about how good AI is at designing and conducting experiments, or whether it’s about how good it is at selling ideas to humans — which we know AI is great at already,” Cook said. “There’s a difference between passing peer review and contributing knowledge to a field.”

Sakana, to its credit, makes no claim that its AI can produce groundbreaking — or even especially novel — scientific work. Rather, the goal of the experiment was to “study the quality of AI-generated research,” the company said, and to highlight the urgent need for “norms regarding AI-generated science.”

“[T]here are difficult questions about whether [AI-generated] science should be judged on its own merits first to avoid bias against it,” the company wrote. “Going forward, we will continue to exchange opinions with the research community on the state of this technology to ensure that it does not develop into a situation in the future where its sole purpose is to pass peer review, thereby substantially undermining the meaning of the scientific peer review process.”

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI 科学研究 同行评审 Sakana
相关文章