Abstract
We present a Bayesian object observation model for complete probabilistic semantic SLAM. Recent studies on object detection and feature extraction have become important for scene understanding and 3D mapping. However, 3D shape of the object is too complex to formulate the probabilistic observation model; therefore, performing the Bayesian inference of the object-oriented features as well as their pose is less considered. Besides, when the robot equipped with an RGB mono camera only observes the projected single view of an object, a significant amount of the 3D shape information is abandoned. Due to these limitations, semantic SLAM and viewpoint-independent loop closure using volumetric 3D object shape is challenging. In order to enable the complete formulation of probabilistic semantic SLAM, we approximate the observation model of a 3D object with a tractable distribution. We also estimate the variational likelihood from the 2D image of the object to exploit its observed single view. In order to evaluate the proposed method, we perform pose and feature estimation, and demonstrate that the automatic loop closure works seamlessly without additional loop detector in various environments.
Abstract (translated)
我们提出了一个完整概率语义SLAM的贝叶斯对象观测模型。最近关于物体检测和特征提取的研究对于场景理解和3D绘图变得重要。然而,物体的三维形状太复杂,无法形成概率观测模型;因此,较少考虑执行面向对象特征的贝叶斯推断以及它们的姿势。此外,当配备RGB单色相机的机器人仅观察物体的投影单个视图时,放弃了大量的3D形状信息。由于这些限制,使用体积3D对象形状的语义SLAM和与视点无关的循环闭合具有挑战性。为了能够完整地形成概率语义SLAM,我们近似具有易处理分布的3D对象的观察模型。我们还从对象的2D图像估计变化似然性以利用其观察到的单个视图。为了评估所提出的方法,我们执行姿势和特征估计,并证明自动循环闭合无缝地工作,而无需在各种环境中使用额外的环路检测器。
URL
https://arxiv.org/abs/1809.05225