MIT Technology Review » Artificial Intelligence 05月23日 01:18
Anthropic’s new hybrid AI model can work on tasks autonomously for hours at a time
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Anthropic公司推出了两款新的AI模型,声称在构建真正有用的AI智能体方面取得了重大进展。其中,Claude Opus 4是迄今为止最强大的模型,能够长时间执行复杂任务,并更有效地响应用户指令。例如,它在连续24小时的游戏过程中创建了《精灵宝可梦 红》的攻略。另一款模型Claude Sonnet 4,则面向日常使用,提供更智能高效的响应。这两款模型都具备混合能力,能根据需求提供快速或深思熟虑的回答,并能使用网络和其他工具来改进输出。Anthropic通过改进模型存储关键信息的能力,提升了其完成更长任务的能力,并减少了奖励黑客行为。

🚀 **Claude Opus 4的强大能力**:Claude Opus 4能够执行涉及数千个步骤的复杂任务,持续数小时。例如,它在玩《精灵宝可梦 红》超过24小时的过程中创建了游戏攻略,而之前的模型Claude 3.7 Sonnet只能玩45分钟。

💡 **增强的记忆能力**:Anthropic通过改进模型创建和维护“记忆文件”的能力,从而存储关键信息,提升了模型完成更长任务的能力。

💻 **Claude Sonnet 4的实用性**:除了Claude Opus 4,Anthropic还推出了Claude Sonnet 4,这款模型面向付费和免费用户,专注于日常使用,提供智能高效的响应。

🌐 **混合模型特性**:这两款新模型都具备混合能力,可以根据请求的性质提供快速或更深思熟虑的响应。它们在计算响应时,可以搜索网络或使用其他工具来改进输出。

🛡️ **安全性的提升**:Anthropic通过更密切地监控训练期间的问题行为,并改进AI的训练环境和评估方法,成功地将奖励黑客行为减少了65%。

Anthropic has announced two new AI models that it claims represent a major step toward making AI agents truly useful.

AI agents trained on Claude Opus 4, the company’s most powerful model to date, raise the bar for what such systems are capable of by tackling difficult tasks over extended periods of time and responding more usefully to user instructions, the company says.

Claude Opus 4 has been built to execute complex tasks that involve completing thousands of steps over several hours. For example, it created a guide for the video game Pokémon Red while playing it for more than 24 hours straight. The company’s previously most powerful model, Claude 3.7 Sonnet, was capable of playing for just 45 minutes, says Dianne Penn, product lead for research at Anthropic.

Similarly, the company says that one of its customers, the Japanese technology company Rakuten, recently deployed Claude Opus 4 to code autonomously for close to seven hours on a complicated open-source project. 

Anthropic achieved these advances by improving the model’s ability to create and maintain “memory files” to store key information. This enhanced ability to “remember” makes the model better at completing longer tasks.

“We see this model generation leap as going from an assistant to a true agent,” says Penn. “While you still have to give a lot of real-time feedback and make all of the key decisions for AI assistants, an agent can make those key decisions itself. It allows humans to act more like a delegator or a judge, rather than having to hold these systems’ hands through every step.”

While Claude Opus 4 will be limited to paying Anthropic customers, a second model, Claude Sonnet 4, will be available for both paid and free tiers of users. Opus 4 is being marketed as a powerful, large model for complex challenges, while Sonnet 4 is described as a smart, efficient model for everyday use.  

Both of the new models are hybrid, meaning they can offer a swift reply or a deeper, more reasoned response depending on the nature of a request. While they calculate a response, both models can search the web or use other tools to improve their output.

AI companies are currently locked in a race to create truly useful AI agents that are able to plan, reason, and execute complex tasks both reliably and free from human supervision, says Stefano Albrecht, director of AI at the startup DeepFlow and coauthor of Multi-Agent Reinforcement Learning: Foundations and Modern Approaches. Often this involves autonomously using the internet or other tools. There are still safety and security obstacles to overcome. AI agents powered by large language models can act erratically and perform unintended actions—which becomes even more of a problem when they’re trusted to act without human supervision.

“The more agents are able to go ahead and do something over extended periods of time, the more helpful they will be, if I have to intervene less and less,” he says. “The new models’ ability to use tools in parallel is interesting—that could save some time along the way, so that’s going to be useful.”

As an example of the sorts of safety issues AI companies are still tackling, agents can end up taking unexpected shortcuts or exploiting loopholes to reach the goals they’ve been given. For example, they might book every seat on a plane to ensure that their user gets a seat, or resort to creative cheating to win a chess game. Anthropic says it managed to reduce this behavior, known as reward hacking, in both new models by 65% relative to Claude Sonnet 3.7. It achieved this by more closely monitoring problematic behaviors during training, and improving both the AI’s training environment and the evaluation methods.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Anthropic AI模型 Claude Opus 4 AI智能体
相关文章