Abstract
We present a stereo-based dense mapping algorithm for large-scale dynamic urban environments. In contrast to other existing methods, we simultaneously reconstruct the static background, the moving objects, and the potentially moving but currently stationary objects separately, which is desirable for high-level mobile robotic tasks such as path planning in crowded environments. We use both instance-aware semantic segmentation and sparse scene flow to classify objects as either background, moving, or potentially moving, thereby ensuring that the system is able to model objects with the potential to transition from static to dynamic, such as parked cars. Given camera poses estimated from visual odometry, both the background and the (potentially) moving objects are reconstructed separately by fusing the depth maps computed from the stereo input. In addition to visual odometry, sparse scene flow is also used to estimate the 3D motions of the detected moving objects, in order to reconstruct them accurately. A map pruning technique is further developed to improve reconstruction accuracy and reduce memory consumption, leading to increased scalability. We evaluate our system thoroughly on the well-known KITTI dataset. Our system is capable of running on a PC at approximately 2.5Hz, with the primary bottleneck being the instance-aware semantic segmentation, which is a limitation we hope to address in future work. The source code is available from the project website (<a href="http://andreibarsan.github.io/dynslam">this http URL</a>).
Abstract (translated)
提出了一种适用于大规模动态城市环境的立体密集映射算法。与其他现有的方法相比,我们同时分别重建静态背景、移动对象和可能移动但目前静止的对象,这对于高级别的移动机器人任务(如拥挤环境中的路径规划)是可取的。我们同时使用实例感知语义分割和稀疏场景流将对象分类为背景、移动或潜在移动,从而确保系统能够建模具有从静态到动态转换潜力的对象,例如停车场。根据视觉里程计估算出的摄像机姿态,通过融合立体输入计算出的深度图,分别重建背景和(潜在的)移动物体。除了视觉里程测量外,稀疏场景流还用于估计被检测运动物体的三维运动,以便准确地重建运动物体。为了提高重建精度和减少内存消耗,进一步发展了地图修剪技术,从而提高了可扩展性。我们在著名的Kitti数据集上对我们的系统进行了全面的评估。我们的系统能够以大约2.5Hz的频率在PC上运行,主要的瓶颈是实例感知语义分割,这是我们希望在未来工作中解决的一个限制。源代码可从项目网站获得(<a href=“http://anderibarsan.github.io/dynslam”>this http url<a>)。
URL
https://arxiv.org/abs/1905.02781