Communications of the ACM - Artificial Intelligence 01月21日
Building Safer and Interoperable AI Systems
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了人工智能(AI)智能体的概念及其发展,回顾了早期知识机器人(knowbots)的概念,并展望了AI智能体在互联网上的交互。文章强调了智能体之间交互的必要性,包括开发标准化的语法和语义,以促进可靠的沟通。同时,文章指出了AI智能体可能存在的风险,例如语言的模糊性导致误解,并介绍了MLCommons AI安全工作组的AILuminate工具,用于测试AI模型的安全性。最后,作者对AI智能体的未来发展保持谨慎乐观,认为在规范的约束下,它们可以高效地完成工作。

💡早期知识机器人(Knowbots)概念的提出:文章回顾了35年前提出的知识机器人概念,这些机器人可以在互联网上移动,执行用户任务,这为今天的AI智能体发展奠定了基础。

🌐AI智能体交互的必要性:为了使AI智能体在互联网上有效工作,需要开发标准化的语法和语义,以促进智能体之间的交流、协议和承诺,并可靠地传递结果。

🛡️AI安全测试的重要性:文章介绍了AILuminate工具,用于通过大量提示测试大型语言模型(LLM)的安全性,并评估其对各种提示的响应,这有助于建立LLM之间的安全指标。

🗣️自然语言的模糊性:文章指出,自然语言的模糊性可能导致AI智能体之间的误解,因此需要控制词汇和语义,以确保交流的清晰和意图的确认。

⚙️AI智能体的应用前景:尽管存在担忧,作者对AI智能体的未来持谨慎乐观态度,认为在适当的约束下,它们可以高效地完成工作,并节省劳动力。

While I am no expert on artificial intelligence (AI), I have some experience with the concept of agents. Thirty-five years ago, my colleague, Robert Kahn, and I explored the idea of knowledge robots (“knowbots” for short)a in the context of digital libraries. In principle, a knowbot was a mobile piece of code that could move around the Internet, landing at servers, where they could execute tasks on behalf of users. The concept is mostly related to finding information and processing it on behalf of a user. We imagined that the knowbot code would land at a serving “knowbot hotel” where it would be given access to content and computing capability. The knowbots would be able to clone themselves to execute their objectives in parallel and would return to their origins bearing the results of their work. Modest prototypes were built in the pre-Web era.

In today’s world, artificially intelligent agents are now contemplated that can interact with each other and with information sources found on the Internet. For this to work, it’s my conjecture that a syntax and semantics will need to be developed and perhaps standardized to facilitate inter-agent interaction, agreements, and commitments for work to be performed, as well as a means for conveying results in reliable and unambiguous ways. A primary question for all such concepts starts with “What could possibly go wrong?”

In the context of AI applications and agents, work is underway to answer that question. I recently found one answer to that in the MLCommons AI Safety Working Group and its tool, AILuminate.b My coarse sense of this is that AILuminate posts a large and widely varying collection of prompts—not unlike the notion of testing software by fuzzingc—looking for inappropriate responses. Large language models (LLMs) can be tested and graded (that’s the hard part) on responses to a wide range of prompts. Some kind of overall safety metric might be established to connect one LLM to another. One might imagine query collections oriented toward exposing particular contextual weaknesses in LLMs. If these ideas prove useful, one could even imagine using them in testing services such as those at Underwriters Laboratory, now called UL Solutions.d UL Solutions already offers software testing among its many other services.

LLMs as agents seem naturally attractive. They can interact via text and speech with humans, so why not with each other? One obvious cautionary note is that people find natural language to be ambiguous, and this can lead to misunderstandings, sometimes serious and sometimes just funny—like giving a flyswatter to someone who asked for a glass of water. Happens to me all the time, but I wear hearing aids, and they don’t always work perfectly! So, I worry about precision and accuracy in inter-agent exchanges. That motivates the possibility of a controlled vocabulary and associated semantics intended to promote clarity and a means for confirming intent in an inter-agent exchange. It is already common for LLMs to generate standardized coded sequences for procedurally calling on other specialized LLMs or applications (for example, mathematical formula manipulators).

Inter-agent exchanges also make me think of sequences of 3D printing steps, where partially printed objects can be fitted into a jig for the next printer to add its step. That’s just an elaboration of the now-classic assembly line concept originated by Henry Ford for producing automobiles. Despite my persistent worry about hallucinating LLMs, colleagues have found their interactive interactions to be generative (no pun intended) and provocative in a kind of mutual brainstorming way. Some scientists are finding these tools to be a stimulus for out-of-the-box thinking.

Despite some trepidation, I am cautiously optimistic that, with some discipline, we may be able to harness these complex creations to carry out useful work in efficient and labor-saving ways. I will stick with my earlier fundamental guiding question though: What could possibly go wrong?

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

人工智能 AI智能体 安全测试 自然语言处理 技术发展
相关文章