MarkTechPost@AI 2024年11月27日
MBA-SLAM: A Novel AI Framework for Robust Dense Visual RGB-D SLAM, Implementing both an Implicit Radiance Fields Version and an Explicit Gaussian Splatting Version
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

MBA-SLAM是一种新颖的AI框架,旨在解决运动模糊图像在稠密视觉SLAM系统中带来的挑战。它通过整合物理运动模糊成像过程到跟踪和建图阶段,有效地重建高质量的稠密3D场景并精确测量相机运动轨迹。该框架采用运动模糊感知跟踪器和基于NeRF或3D高斯散射的去模糊映射器,显著提高了跟踪精度和图像重建质量,并在ScanNet和TUM RGB-D等数据集上表现出色。与现有方法相比,MBA-SLAM在跟踪精度、图像质量和处理速度方面均有提升,展现出在动态环境下SLAM应用的巨大潜力。

🤔 **运动模糊对稠密视觉SLAM的影响:** 运动模糊导致相机姿态估计不准确和多视角几何不一致,从而影响3D地图重建质量。

🚀 **MBA-SLAM框架的核心组件:** 该框架整合了运动模糊感知跟踪器和基于NeRF或3D高斯散射的去模糊映射器,分别用于处理相机运动和3D场景重建。

📊 **实验结果表明:** MBA-SLAM在ScanNet和TUM RGB-D数据集上均取得了优异的跟踪精度,并在图像重建质量和处理速度方面超越了ORB-SLAM3、DSO等现有方法。

💡 **场景表示方法:** MBA-SLAM探索了两种场景表示方法:隐式神经辐射场(NeRF)和显式3D高斯散射(3D-GS),前者帧率更高但渲染质量较低,后者质量更好但帧率较低。

⏱️ **实时处理与加速:** 通过CUDA加速,MBA-SLAM实现了实时处理,速度是其他方法的5倍。

SLAM (Simultaneous Localization and Mapping) is one of the important techniques used in robotics and computer vision. It helps machines understand where they are and create a map of their surroundings. Motion-blurred images face difficulties in dense visual SLAM systems for two reasons: 1) Inaccurate pose estimation during tracking: Current photo-realistic dense visual SLAM algorithms rely on clear images to estimate camera positions by ensuring consistent brightness across views. This affects the mapping process, leading to inconsistent multi-view geometry. 2)  Inconsistent multi-view geometry in mapping: Poor image quality from various views may lead to incorrect features, which cause errors in 3D geometry and a low-quality reconstruction of the 3D map. Combining these two factors, existing dense virtual SLAM systems would usually perform poorly when handling motion-blurred images. 

Traditional sparse SLAM methods use sparse point clouds for map reconstruction. Recent learning-based dense SLAM systems focus on generating dense maps important for downstream tasks. Neural Radiance Fields (NeRF) and 3D Gaussian Splatting (3DGS) have been used with SLAM systems to create realistic 3D scenes, improving map quality and texture. However, existing methods heavily rely on high-quality, sharp RGB-D inputs, which pose challenges when dealing with motion-blurred frames, often encountered in low-light or long-exposure conditions, which result in low precision and efficiency of localization and mapping in various methods.

To solve these problems, a group of researchers from China conducted detailed research and proposed MBA-SLAM, a photo-realistic dense RGB-D SLAM pipeline designed to handle motion-blurred inputs effectively. This approach integrates the physical motion blur imaging process into the tracking and mapping stages. The main objective of this framework is to reconstruct high-quality, dense 3D scenes and accurately measure camera motion trajectories, which was achieved by integrating two key components: a motion blur-aware tracker and a bundle-adjusted deblur mapper based on NeRF or 3D Gaussian Splatting.

The method used a continuous motion model to track the camera’s movement during exposure. The system considered the camera’s start and end positions for each motion-blurred image. In tracking, a sharp reference image was rendered, blurred to match the current image, and compared to improve the motion estimate. The camera trajectories and 3D scenes were optimized in mapping to reduce image-matching errors. Two scene representations were explored: implicit neural radiance fields (NeRF) and explicit 3D Gaussian Splatting (3D-GS). NeRF achieved higher frame rates but lower rendering quality, while 3D-GS offered better quality at the cost of lower frame rates. 

The method showed a measure reduction in tracking errors, with the ScanNet dataset yielding an ATE RMSE of 0.053, outperforming ORB-SLAM3 (0.081) and LDS-SLAM (0.071). On the TUM RGB-D dataset, MBA-SLAM achieved an ATE RMSE of 0.062, showing its superior tracking precision. In image reconstruction, MBA-SLAM excelled with a PSNR of 31.2 dB on the ArchViz dataset and an SSIM of 0.96 on ScanNet, outperforming methods like ORB-SLAM3 and DSO in terms of quality. The LPIPS score of MBA-SLAM is also reported to be 0.18, which reflects better perceptual quality. Radiance fields and Gaussian splatting improved image quality, while CUDA acceleration enabled real-time processing, making it 5 times faster than others. MBA SLAM provided improved accuracy in tracking, better image quality, and speed compared to others, and it seemed to promise an application in SLAM scenarios with motion blur due to dynamism in the environment.

In summary, the proposed framework MBA-SLAM effectively addresses problems in the SLAM system. With its physical motion blur image formation model, highly CUDA-optimized blur-aware tracker, and deblurring mapper, the MBA-SLAM tracked accurate camera motion trajectories within exposure time and reconstructed a sharp and photo-realistic map for the given video sequence input. It performed much better than the previous methods on existing and real-world datasets. This work marks a significant development in the field of SLAM systems and can be used as a baseline for future advancement and research!


Check out the Paper and GitHub. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter.. Don’t Forget to join our 55k+ ML SubReddit.

Evaluation of Large Language Model Vulnerabilities: A Comparative Analysis of Red Teaming Techniques’ Read the Full Report (Promoted)

The post MBA-SLAM: A Novel AI Framework for Robust Dense Visual RGB-D SLAM, Implementing both an Implicit Radiance Fields Version and an Explicit Gaussian Splatting Version appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

SLAM 运动模糊 稠密视觉 NeRF 3D高斯散射
相关文章