Nvidia Blog 前天 19:05
NVIDIA Research Casts New Light on Scenes With AI-Powered Rendering for Physical AI Development
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

英伟达研究团队开发出名为DiffusionRenderer的AI技术,能够对视频光照进行精细控制,实现白天变黑夜、晴天转阴天等效果。该技术融合了逆向渲染和正向渲染,性能优于现有方法,为创意产业和物理AI发展提供了强大工具。DiffusionRenderer可用于广告、电影和游戏开发,以及自动驾驶汽车模型的训练,通过增强合成数据集,提升模型对各种光照条件的适应能力。该技术是英伟达在计算机视觉与模式识别会议(CVPR)上展示的众多研究成果之一。

💡 DiffusionRenderer是一种用于视频光照控制、编辑和合成数据增强的新技术,它通过AI近似模拟真实世界的光照行为。

☀️ 该技术能够实现将白天场景转换为夜景、将晴朗的下午转换为阴天、并将刺眼的荧光灯光转换为柔和的自然光照效果。

🎬 DiffusionRenderer将传统的逆向渲染和正向渲染结合在一个统一的神经渲染引擎中,能够从单个2D视频数据中估计法线、金属度和粗糙度等属性,进而生成新的阴影和反射,改变光源,编辑材质,并在场景中插入新对象。

🚗 自动驾驶汽车开发者可以使用DiffusionRenderer增强数据集,模拟各种光照条件,从而训练出能够应对复杂光照环境的模型。

✨ DiffusionRenderer已与Cosmos Predict-1集成,后者是一个用于生成逼真、物理感知未来世界状态的世界基础模型。这种集成提高了DiffusionRenderer的去光照和重新光照质量,使其结果更加清晰、准确和时间一致。

NVIDIA Research has developed an AI light switch for videos that can turn daytime scenes into nightscapes, transform sunny afternoons to cloudy days and tone down harsh fluorescent lighting into soft, natural illumination.

Called DiffusionRenderer, it’s a new technique for neural rendering — a process that uses AI to approximate how light behaves in the real world. It brings together two traditionally distinct processes — inverse rendering and forward rendering — in a unified neural rendering engine that outperforms state-of-the-art methods.

DiffusionRenderer provides a framework for video lighting control, editing and synthetic data augmentation, making it a powerful tool for creative industries and physical AI development.

Creators in advertising, film and game development could use applications based on DiffusionRenderer to add, remove and edit lighting in real-world or AI-generated videos. Physical AI developers could use it to augment synthetic datasets with a greater diversity of lighting conditions to train models for robotics and autonomous vehicles (AVs).

DiffusionRenderer is one of over 60 NVIDIA papers accepted to the Computer Vision and Pattern Recognition (CVPR) conference, taking place June 11-15 in Nashville, Tennessee.

Creating AI That Delights

DiffusionRenderer tackles the challenge of de-lighting and relighting a scene from only 2D video data.

De-lighting is a process that takes an image and removes its lighting effects, so that only the underlying object geometry and material properties remain. Relighting does the opposite, adding or editing light in a scene while maintaining the realism of complex properties like object transparency and specularity — how a surface reflects light.

Classic, physically based rendering pipelines need 3D geometry data to calculate light in a scene for de-lighting and relighting. DiffusionRenderer instead uses AI to estimate properties including normals, metallicity and roughness from a single 2D video.

With these calculations, DiffusionRenderer can generate new shadows and reflections, change light sources, edit materials and insert new objects into a scene — all while maintaining realistic lighting conditions.

Using an application powered by DiffusionRenderer, AV developers could take a dataset of mostly daytime driving footage and randomize the lighting of every video clip to create more clips representing cloudy or rainy days, evenings with harsh lighting and shadows, and nighttime scenes. With this augmented data, developers can boost their development pipelines to train, test and validate AV models that are better equipped to handle challenging lighting conditions.

Creators who capture content for digital character creation or special effects could use DiffusionRenderer to power a tool for early ideation and mockups — enabling them to explore and iterate through various lighting options before moving to expensive, specialized light stage systems to capture production-quality footage.

Enhancing DiffusionRenderer With NVIDIA Cosmos

Since completing the original paper, the research team behind DiffusionRenderer has integrated their method with Cosmos Predict-1, a suite of world foundation models for generating realistic, physics-aware future world states.

By doing so, the researchers observed a scaling effect, where applying Cosmos Predict’s larger, more powerful video diffusion model boosted the quality of DiffusionRenderer’s de-lighting and relighting correspondingly — enabling sharper, more accurate and temporally consistent results.

Cosmos Predict is part of NVIDIA Cosmos, a platform of world foundation models, tokenizers, guardrails and an accelerated data processing and curation pipeline to accelerate synthetic data generation for physical AI development. Read about the new Cosmos Predict-2 model on the NVIDIA Technical Blog.

NVIDIA Research at CVPR 

At CVPR, NVIDIA researchers are presenting dozens of papers on topics spanning automotive, healthcare, robotics and more. Three NVIDIA papers are nominated for this year’s Best Paper Award:

NVIDIA was also named an Autonomous Grand Challenge winner at CVPR, marking the second consecutive year NVIDIA topped the leaderboard in the end-to-end category — and the third consecutive year winning an Autonomous Grand Challenge award at the conference.

Learn more about NVIDIA Research, a global team of hundreds of scientists and engineers focused on topics including AI, computer graphics, computer vision, self-driving cars and robotics.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

DiffusionRenderer AI 视频光照 英伟达 CVPR
相关文章