MIT News - Machine learning 01月30日
MIT students' works redefine human-AI collaboration
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

MIT的“互动智能”课程在NeurIPS大会上展示了一系列创新项目,探索人工智能在物理世界的应用。这些项目涵盖舞蹈、音乐、故事叙述等领域,旨在通过人机协作激发创造力,重塑教育和社会互动。其中,“Be the Beat”将舞蹈与音乐结合,让舞者通过肢体动作创作音乐;“A Mystery for You”是一款培养批判性思维的游戏,让玩家充当事实核查员;“Memorscope”则创造了集体回忆,通过人脸互动探索共享记忆;“Narratron”通过影子戏和语言模型共同创作儿童故事;“Perfect Syntax”则通过AI操纵视频片段,探索运动和时间的语法逻辑。这些项目展现了AI超越自动化,促进创新和人机协作的潜力。

🎶 “Be the Beat”:这款AI驱动的音箱能根据舞者的动作推荐音乐,打破了传统音乐引导舞蹈的模式,让舞蹈成为音乐创作的源泉,通过PoseNet分析舞姿,再用大型语言模型匹配音乐风格。

🤔 “A Mystery for You”:这款教育游戏利用LLM和实体界面,让玩家扮演事实核查员,通过分析AI生成的新闻,培养批判性思维和信息辨别能力,采用实体卡片替代触屏,鼓励更深入的思考。

👁️‍🗨️ “Memorscope”:这款设备通过人脸互动创造集体回忆,利用AI模型对记忆进行美学和情感上的诠释,使记忆从静态照片转变为动态、交互式的叙事空间,超越传统共享相册的限制。

🎭 “Narratron”:这是一个互动投影仪,通过影子戏和大型语言模型共同创作儿童故事,用户可以通过手影定义角色,系统会实时生成故事情节,结合视觉、听觉和身体互动。

🎬 “Perfect Syntax”:这是一个视频艺术作品,通过AI操纵视频片段,探索运动和时间的语法逻辑,思考机器如何理解和表现时间和运动的流逝,引发对感知、技术和运动之间关系的思考。

Imagine a boombox that tracks your every move and suggests music to match your personal dance style. That’s the idea behind “Be the Beat,” one of several projects from MIT course 4.043/4.044 (Interaction Intelligence), taught by Marcelo Coelho in the Department of Architecture, that were presented at the 38th annual NeurIPS (Neural Information Processing Systems) conference in December 2024. With over 16,000 attendees converging in Vancouver, NeurIPS is a competitive and prestigious conference dedicated to research and science in the field of artificial intelligence and machine learning, and a premier venue for showcasing cutting-edge developments.

The course investigates the emerging field of large language objects, and how artificial intelligence can be extended into the physical world. While “Be the Beat” transforms the creative possibilities of dance, other student submissions span disciplines such as music, storytelling, critical thinking, and memory, creating generative experiences and new forms of human-computer interaction. Taken together, these projects illustrate a broader vision for artificial intelligence: one that goes beyond automation to catalyze creativity, reshape education, and reimagine social interactions.

Be the Beat 

“Be the Beat,” by Ethan Chang, an MIT mechanical engineering and design student, and Zhixing Chen, an MIT mechanical engineering and music student, is an AI-powered boombox that suggests music from a dancer's movement. Dance has traditionally been guided by music throughout history and across cultures, yet the concept of dancing to create music is rarely explored.

“Be the Beat” creates a space for human-AI collaboration on freestyle dance, empowering dancers to rethink the traditional dynamic between dance and music. It uses PoseNet to describe movements for a large language model, enabling it to analyze dance style and query APIs to find music with similar style, energy, and tempo. Dancers interacting with the boombox reported having more control over artistic expression and described the boombox as a novel approach to discovering dance genres and choreographing creatively.

A Mystery for You

“A Mystery for You,” by Mrinalini Singha SM ’24, a recent graduate in the Art, Culture, and Technology program, and Haoheng Tang, a recent graduate of the Harvard University Graduate School of Design, is an educational game designed to cultivate critical thinking and fact-checking skills in young learners. The game leverages a large language model (LLM) and a tangible interface to create an immersive investigative experience. Players act as citizen fact-checkers, responding to AI-generated “news alerts” printed by the game interface. By inserting cartridge combinations to prompt follow-up “news updates,” they navigate ambiguous scenarios, analyze evidence, and weigh conflicting information to make informed decisions.

This human-computer interaction experience challenges our news-consumption habits by eliminating touchscreen interfaces, replacing perpetual scrolling and skim-reading with a haptically rich analog device. By combining the affordances of slow media with new generative media, the game promotes thoughtful, embodied interactions while equipping players to better understand and challenge today’s polarized media landscape, where misinformation and manipulative narratives thrive.

Memorscope

“Memorscope,” by MIT Media Lab research collaborator Keunwook Kim, is a device that creates collective memories by merging the deeply human experience of face-to-face interaction with advanced AI technologies. Inspired by how we use microscopes and telescopes to examine and uncover hidden and invisible details, Memorscope allows two users to “look into” each other’s faces, using this intimate interaction as a gateway to the creation and exploration of their shared memories.

The device leverages AI models such as OpenAI and Midjourney, introducing different aesthetic and emotional interpretations, which results in a dynamic and collective memory space. This space transcends the limitations of traditional shared albums, offering a fluid, interactive environment where memories are not just static snapshots but living, evolving narratives, shaped by the ongoing relationship between users.

Narratron

“Narratron,” by Harvard Graduate School of Design students Xiying (Aria) Bao and Yubo Zhao, is an interactive projector that co-creates and co-performs children's stories through shadow puppetry using large language models. Users can press the shutter to “capture” protagonists they want to be in the story, and it takes hand shadows (such as animal shapes) as input for the main characters. The system then develops the story plot as new shadow characters are introduced. The story appears through a projector as a backdrop for shadow puppetry while being narrated through a speaker as users turn a crank to “play” in real time. By combining visual, auditory, and bodily interactions in one system, the project aims to spark creativity in shadow play storytelling and enable multi-modal human-AI collaboration.

Perfect Syntax

“Perfect Syntax,” by Karyn Nakamura ’24, is a video art piece examining the syntactic logic behind motion and video. Using AI to manipulate video fragments, the project explores how the fluidity of motion and time can be simulated and reconstructed by machines. Drawing inspiration from both philosophical inquiry and artistic practice, Nakamura's work interrogates the relationship between perception, technology, and the movement that shapes our experience of the world. By reimagining video through computational processes, Nakamura investigates the complexities of how machines understand and represent the passage of time and motion.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

人工智能 人机交互 创新项目 MIT NeurIPS
相关文章