All Content from Business Insider 14小时前
OpenAI and Microsoft are dueling over AGI. These real-world tests will prove when AI is really better than humans.
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了OpenAI与微软合作关系中人工智能通用智能(AGI)的重要性。文章指出,微软从OpenAI的合作中获得巨大利益,而AGI的实现将改变这种关系。作者提出了几个现实世界的测试,以验证AGI是否真正实现,例如AI能否有效处理邮件、避免特斯拉汽车的颠簸、以及预测未来。文章还引用了专家观点,认为当前的AI模型仍缺乏从经验中学习的核心算法,距离真正的AGI可能还很遥远。

🤖 微软与OpenAI的合作关系高度依赖于AGI的实现。在AGI达成之前,微软将从OpenAI获得重要的技术和经济利益,包括OpenAI需要与微软分享其收入的很大一部分,这涉及数十亿美元的资金。

🤔 作者提出了多个现实世界的测试来检验AGI的实际应用。这些测试包括:OpenAI和Anthropic的公关部门能否使用AI回答记者提问;微软的Outlook邮件系统能否智能地过滤垃圾邮件,同时保留重要邮件;AI能否阻止Cactus Warehouse频繁发送促销短信。

🚗 作者还提到了特斯拉汽车的自动驾驶系统(FSD)在行驶中避开路面坑洼的能力,以及AI模型预测未来的能力。这些测试旨在揭示AI在解决实际问题上的局限性。

🏀 文章通过谷歌Gemini的广告案例,强调了AI在实际操作中的不足。虽然AI可以提供组装篮球网的说明,但它无法像人类一样直接完成物理组装工作。

💡 专家Konstantin Mishchenko认为,当前的LLMs(大型语言模型)只是模仿互联网数据,缺乏从经验中直接获取智能的核心算法,这表明LLMs与真正的智能之间可能存在巨大差距。

OpenAI CEO Sam Altman (left) and Microsoft CEO Satya Nadella

AGI is a pretty silly debate. It's only really important in one way: It governs how the world's most important AI partnership will change in the coming months. That's the deal between OpenAI and Microsoft.

This is the situation right now: Until OpenAI achieves Artificial General Intelligence — where AI capabilities surpass those of humans — Microsoft gets a lot of valuable technological and financial benefits from the startup. For instance, OpenAI must share a significant portion of its revenue with Microsoft. That's billions of dollars.

One could reasonably argue that this might be why Sam Altman bangs on about OpenAI getting close to AGI soon.

Many other experts in the AI field don't talk about this much, or they think the AGI debate is off base in various ways, or just not that important. Even Anthropic CEO Dario Amodei, one of the biggest AI boosters on the planet, doesn't like to talk about AGI.

Microsoft CEO Satya Nadella sees things very differently. Wouldn't you? If another company is contractually required to give you oodles of money if they don't reach AGI, then you're probably not going to think we're close to AGI!

Nadella has called the push toward AGI "benchmark hacking," which is so delicious. This refers to AI researchers and labs designing AI models to perform well on wonky industry benchmarks, rather than in real life.

Here's OpenAI's official definition of AGI: "highly autonomous systems that outperform humans at most economically valuable work."

Other experts have defined it slightly differently. But the main point is that AI machines and software must be better than humans at a wide variety of useful tasks. You can already train an AI model to be better at one or two specific things, but to get to artificial general intelligence, machines must be able to do many different things better than humans.

My real-world AGI tests

Over the past few months, I've devised several real-world tests to see if we've reached AGI. These are fun or annoying everyday things that should just work in a world of AGI, but they don't right now for me. I also canvassed input from readers of my Tech Memo newsletter and tapped my source network for fun suggestions.

Here are my real-world tests that will prove we've reached AGI:

Yes, I know these tests seem a bit silly — but AI benchmarks are not the real world, and they can be pretty easily gamed.

That last basketball net test is particularly telling for me. Getting an AI system and software to actually assemble a basketball net — that might happen sometime soon. But, getting the same system to do a lot of other physical-world manipulation stuff better than humans, too? Very hard and probably not possible for a very long time.

As OpenAI and Microsoft try to resolve their differences, the companies can tap experts to weigh in on whether the startup has reached AGI or not, per the terms of their existing contract, according to The Information. I'm happy to be an expert advisor here. Sam and Satya, let me know if you want help!

For now, I'll leave the final words to a real AI expert. Konstantin Mishchenko, an AI research scientist at Meta, recently tweeted this, while citing a blog by another respected expert in the field, Sergey Levine:

"While LLMs learned to mimic intelligence from internet data, they never had to actually live and acquire that intelligence directly. They lack the core algorithm for learning from experience. They need a human to do that work for them," Mishchenko wrote, referring to AI models known as large language models.

"This suggests, at least to me, that the gap between LLMs and genuine intelligence might be wider than we think. Despite all the talk about AGI either being already here or coming next year, I can't shake off the feeling it's not possible until we come up with something better than a language model mimicking our own idea of how an AI should look," he concluded.

Sign up for BI's Tech Memo newsletter here. Reach out to me via email at abarr@businessinsider.com.

Read the original article on Business Insider

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AGI OpenAI 微软 AI测试 合作关系
相关文章