ΑΙhub 05月14日 19:19
Robot see, robot do: System learns after watching how-tos
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

康奈尔大学的研究人员开发了一种名为RHyME的新型机器人框架,它利用人工智能使机器人能够通过观看单个操作视频来学习任务。RHyME通过显著减少训练机器人所需的时间、精力和金钱,加速机器人系统的开发和部署。该系统通过存储之前的例子并借鉴类似动作的视频,使机器人能够适应人类和机器人动作之间的差异,仅需30分钟的机器人数据,任务成功率提高了50%以上,从而更有效地训练机器人。

🤖RHyME(Retrieval for Hybrid Imitation under Mismatched Execution)是一个AI驱动的机器人框架,让机器人通过观看单个教学视频学习任务,旨在减少机器人训练所需的时间、精力和资金。

🧠RHyME的核心思想是让机器人存储之前的任务示例,并在执行新任务时,通过借鉴和连接类似动作的视频来完成学习。这使得机器人能够适应人类和机器人动作之间的差异。

⏱️RHyME显著降低了机器人训练所需的数据量。研究人员声称,使用RHyME仅需30分钟的机器人数据,在实验室环境中,使用该系统训练的机器人任务成功率比以前的方法提高了50%以上。

🔄RHyME的原理类似于翻译,即将人类的任务指令“翻译”成机器人可以执行的动作。它通过检索和模仿人类的动作,使机器人能够更快地适应真实世界的环境。

Kushal Kedia (left) and Prithwish Dan (right) are members of the development team behind RHyME, a system that allows robots to learn tasks by watching a single how-to video.

By Louis DiPietro

Cornell researchers have developed a new robotic framework powered by artificial intelligence – called RHyME (Retrieval for Hybrid Imitation under Mismatched Execution) – that allows robots to learn tasks by watching a single how-to video. RHyME could fast-track the development and deployment of robotic systems by significantly reducing the time, energy and money needed to train them, the researchers said.

“One of the annoying things about working with robots is collecting so much data on the robot doing different tasks,” said Kushal Kedia, a doctoral student in the field of computer science and lead author of a corresponding paper on RHyME. “That’s not how humans do tasks. We look at other people as inspiration.”

Kedia will present the paper, One-Shot Imitation under Mismatched Execution, in May at the Institute of Electrical and Electronics Engineers’ International Conference on Robotics and Automation, in Atlanta.

Home robot assistants are still a long way off – it is a very difficult task to train robots to deal with all the potential scenarios that they could encounter in the real world. To get robots up to speed, researchers like Kedia are training them with what amounts to how-to videos – human demonstrations of various tasks in a lab setting. The hope with this approach, a branch of machine learning called “imitation learning,” is that robots will learn a sequence of tasks faster and be able to adapt to real-world environments.

“Our work is like translating French to English – we’re translating any given task from human to robot,” said senior author Sanjiban Choudhury, assistant professor of computer science in the Cornell Ann S. Bowers College of Computing and Information Science.

This translation task still faces a broader challenge, however: Humans move too fluidly for a robot to track and mimic, and training robots with video requires gobs of it. Further, video demonstrations – of, say, picking up a napkin or stacking dinner plates – must be performed slowly and flawlessly, since any mismatch in actions between the video and the robot has historically spelled doom for robot learning, the researchers said.

“If a human moves in a way that’s any different from how a robot moves, the method immediately falls apart,” Choudhury said. “Our thinking was, ‘Can we find a principled way to deal with this mismatch between how humans and robots do tasks?’”

RHyME is the team’s answer – a scalable approach that makes robots less finicky and more adaptive. It trains a robotic system to store previous examples in its memory bank and connect the dots when performing tasks it has viewed only once by drawing on videos it has seen. For example, a RHyME-equipped robot shown a video of a human fetching a mug from the counter and placing it in a nearby sink will comb its bank of videos and draw inspiration from similar actions – like grasping a cup and lowering a utensil.

RHyME paves the way for robots to learn multiple-step sequences while significantly lowering the amount of robot data needed for training, the researchers said. They claim that RHyME requires just 30 minutes of robot data; in a lab setting, robots trained using the system achieved a more than 50% increase in task success compared to previous methods.

“This work is a departure from how robots are programmed today. The status quo of programming robots is thousands of hours of tele-operation to teach the robot how to do tasks. That’s just impossible,” Choudhury said. “With RHyME, we’re moving away from that and learning to train robots in a more scalable way.”

This research was supported by Google, OpenAI, the U.S. Office of Naval Research and the National Science Foundation.

Read the work in full

One-Shot Imitation under Mismatched Execution, Kushal Kedia, Prithwish Dan, Angela Chao, Maximus Adrian Pace, Sanjiban Choudhury.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

RHyME 机器人 人工智能 模仿学习
相关文章