论文标题
Dymslam:基于几何运动分段的4D动态场景重建
DymSLAM:4D Dynamic Scene Reconstruction Based on Geometrical Motion Segmentation
论文作者
论文摘要
大多数猛击算法基于以下假设:场景是静态的。但是,实际上,大多数场景都是动态的,通常包含移动对象,这些方法不合适。在本文中,我们介绍了Dymslam,这是一个动态的立体声视觉猛击系统,能够用刚体移动的对象重建4D(3D +时间)动态场景。 Dymslam的唯一输入是立体视频,其输出包括静态环境的密集图,移动对象的3D模型以及相机的轨迹和移动对象。我们首先通过使用传统的大满贯方法来检测并匹配连续帧之间的有趣点。那么属于不同运动模型的有趣点(包括刚体运动对象的自我运动和运动模型)将通过多模型拟合方法进行分割。基于属于自我运动的有趣点,我们能够估计相机的轨迹并重建静态背景。然后使用属于刚性运动对象的运动模型的有趣点来估算其相对运动模型并重建对象的3D模型。然后,我们将相对运动转换为全局参考框架中移动对象的轨迹。最后,我们通过考虑其运动轨迹以获得4D(3D+时间)序列,将移动对象的3D模型融合到环境的3D图中。 Dymslam获得了有关动态对象的信息,而不是忽略它们,适合未知的刚性对象。因此,提出的系统允许机器人用于高级任务,例如避免动态对象的障碍物。我们在现实世界环境中进行了实验,在现实世界中,相机和物体都在广泛的范围内移动。
Most SLAM algorithms are based on the assumption that the scene is static. However, in practice, most scenes are dynamic which usually contains moving objects, these methods are not suitable. In this paper, we introduce DymSLAM, a dynamic stereo visual SLAM system being capable of reconstructing a 4D (3D + time) dynamic scene with rigid moving objects. The only input of DymSLAM is stereo video, and its output includes a dense map of the static environment, 3D model of the moving objects and the trajectories of the camera and the moving objects. We at first detect and match the interesting points between successive frames by using traditional SLAM methods. Then the interesting points belonging to different motion models (including ego-motion and motion models of rigid moving objects) are segmented by a multi-model fitting approach. Based on the interesting points belonging to the ego-motion, we are able to estimate the trajectory of the camera and reconstruct the static background. The interesting points belonging to the motion models of rigid moving objects are then used to estimate their relative motion models to the camera and reconstruct the 3D models of the objects. We then transform the relative motion to the trajectories of the moving objects in the global reference frame. Finally, we then fuse the 3D models of the moving objects into the 3D map of the environment by considering their motion trajectories to obtain a 4D (3D+time) sequence. DymSLAM obtains information about the dynamic objects instead of ignoring them and is suitable for unknown rigid objects. Hence, the proposed system allows the robot to be employed for high-level tasks, such as obstacle avoidance for dynamic objects. We conducted experiments in a real-world environment where both the camera and the objects were moving in a wide range.