少点错误 03月28日 22:12
Will the Need to Retrain AI Models from Scratch Block a Software Intelligence Explosion?
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了AI完全自动化AI研发后,软件智能是否会快速发展,即“软件智能爆炸”(SIE)。文章关注了重新训练AI系统对SIE的影响。研究表明,虽然重新训练会稍微减缓软件进步的速度,但不会阻止SIE的发生。然而,如果训练时间较长,或加速过程过于激进,重新训练的影响会更显著。文章通过理论分析和简单模型,量化了重新训练对SIE的影响,并指出,除非训练时间大幅缩短或效率提升,否则SIE不太可能在10个月内完成。

🤔 重新训练不会阻止软件进步加速。文章分析认为,即使考虑到重新训练的需求,软件进步仍将加速,不会阻碍SIE的发生。

⏳ 重新训练会稍微减缓软件进步的速度。理论分析表明,重新训练会使软件进步的加速过程略有减缓,但这种减缓的程度相对较小,例如,加速时间可能增加约20%。

⏱️ 训练时间是关键因素。文章指出,如果初始训练时间较长(如100天),重新训练会导致SIE所需时间增加约3倍。相反,如果训练时间较短(如30天),影响会减小,SIE所需时间增加约2倍。

🚀 效率提升和优化也至关重要。除了训练时间,运行时效率的提高和其他训练后增强也可能加速SIE,即使不从头开始重新训练。

Published on March 28, 2025 2:12 PM GMT

Tl;dr: no.

This is a rough research note – we’re sharing it for feedback and to spark discussion. We’re less confident in its methods and conclusions.

Once AI fully automates AI R&D, there might be a period of fast and accelerating software progress – a software intelligence explosion (SIE).

One objection to this is that it takes a long time to train SOTA AI systems from scratch. Would retraining each new generation of AIs stop progress accelerating during an SIE? If not, would it significantly delay an SIE? This post investigates this objection.

Here are my tentative bottom lines:

How did I reach these tentative conclusions?

Theoretical analysis
First, I conducted a very basic theoretical analysis on the effect of retraining. I took a standard semi-endogenous growth model of tech development, and used empirical estimates for the diminishing returns to software progress. This is the simplest model I know of to estimate the dynamics of an SIE – and by default it doesn’t account for retraining.

To understand how fast software progress1 accelerates, we can ask: how many times must software double before the pace of software progress doubles? This is a measure of how quickly software progress accelerates: lower numbers mean faster acceleration.

Without including retraining, my median parameters imply that once AI R&D is fully automated, software must double ~five times before the pace of software progress doubles. (There’s large uncertainty.)2 So progress accelerates gradually.

Accounting for retraining increases the number from five to six. Training runs get progressively shorter over time, and the SIE still accelerates but slightly more slowly. (See below for more explanation.)

In a very aggressive scenario where software must only double once before the pace of progress doubles, retraining makes a big difference by increasing this to twice. So retraining makes a bigger difference to these aggressive scenarios, making them significantly less extreme.

Simple spreadsheet models
Second, I made very simple spreadsheet models of an SIE – again based on semi-endogenous growth models – one without retraining and one with retraining. Both sheets use the same parameters (other than whether to include retraining) and both calculate the time between the AI R&D being automated and AI capabilities going to infinity. I assumed that AI algorithms become 2X as efficient every month – that’s about ~10X faster progress than today.

Results:

(These slowdowns are longer than the slowdown predicted by the theoretical analysis because the theoretical analysis assumes that training runs get gradually shorter as the SIE gets closer, so are already very short at the point at which the spreadsheet model begins, and so have less of a slowing effect. See below for details.)

Full post.

Twitter thread.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

人工智能 AI研发 软件智能爆炸 重新训练 技术进步
相关文章