少点错误 2024年09月15日
OpenAI o1, Llama 4, and AlphaZero of LLMs
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了 GPT-4 级模型的相关情况,包括其当前的能力表现、后训练的方法及可能的发展,还提到了模型改进对其能力的提升以及面临的问题。

🎯GPT-4 级开放权重模型如 Llama-3-405B 似乎不具备危险认知能力,OpenAI o1 表明该级别模型可通过后训练产生有用的长时程推理痕迹。

📊后训练目前是使用合成数据和人类标记数据的技术组合,人类标记数据虽能显著提高质量,但收集缓慢且扩展性差,合成数据在后期训练中愈发重要且易于扩展。

💡GPT-4 级语言模型在多数情况下能清晰通过阅读理解,OpenAI o1 在此基础上进一步提升。若下一级规模的预训练模型如 Llama 4 以开放权重形式提供,可能最初与当前模型表现相似,但后训练能显著提升其能力。

🔍当前存在的问题是新一级的预训练规模即将到来,而廉价应用长时程推理后训练的能力可能随后跟进,但目前对其最终能力水平尚不明确。

Published on September 14, 2024 9:27 PM GMT

GPT-4 level open weights models like Llama-3-405B don't seem capable of dangerous cognition. OpenAI o1 demonstrates that a GPT-4 level model can be post-trained into producing useful long horizon reasoning traces. AlphaZero shows how capabilities can be obtained from compute alone, with no additional data. If there is a way of bringing these together, the apparent helplessness of the current generation of open weights models might prove misleading.

Post-training is currently a combination of techniques that use synthetic data and human labeled data. Human labeled data significantly improves quality, but its collection is slow and scales poorly. Synthetic data is an increasingly useful aspect of post-training, and automated aspects of its generation scale easily. Unlike weaker models, GPT-4 level LLMs clearly pass reading comprehension on most occasions, OpenAI o1 improves on this further. This suggests that at some point human data might become mostly unnecessary in post-training, even if it still slightly helps. Without it, post-training becomes automated and gets to use more compute, while avoiding the need for costly and complicated human labeling.

A pretrained model at the next level of scale, such as Llama 4, if made available in open weights, might initially look approximately as tame as current models. OpenAI o1 demonstrates that useful post-training for long sequences of System 2 reasoning is possible. In the case of o1 in particular, this might involve a lot of human labeling, making its reproduction a very complicated process (at least if the relevant datasets are not released, and the reasoning traces themselves are not leaked in large quantities). But if some generally available chatbots at the next level of scale are good enough at automating labeling, this complication could be sidestepped, with o1 style post-training cheaply reproduced on top of a previously released open weights model.

So there is an overhang in an open weights model that's distributed without long horizon reasoning post-training, since applying such post-training significantly improves its capabilities, making perception of its prior capabilities inadequate. The problem right now is that a new level of pretraining scale is approaching in the coming months, while ability to cheaply apply long horizon reasoning post-training might follow shortly thereafter, possibly unlocked by these very same models at the new level of pretraining scale (since it might currently be too expensive for most actors to implement, or to do enough experiments to figure out how). The resulting level of capabilities is currently unknown, and could well remain unknown outside the leading labs until after the enabling artifacts of the open weights pretrained models at the next level of scale have already been published.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

GPT-4 级模型 后训练 语言模型 能力提升 未知能力
相关文章