论文标题
3D对象重建和6D置式估计来自2D形状,用于机器人抓紧物体
3D object reconstruction and 6D-pose estimation from 2D shape for robotic grasping of objects
论文作者
论文摘要
我们从2D图像中提出了一种3D对象重建和6D置式估计的方法,该方法使用有关对象形状作为主要键的知识。在拟议的管道中,2D图像中对象的识别和标记传递了2D段剪影,这些轮廓与从代表公认对象类的3D模型的各种视图获得的投影的2D剪影进行了比较。通过直接从2D图像计算转换参数,减少了注册过程中所需的自由参数数量,从而使该方法可行。此外,使用校准的设置,还采用了3D转换和投影几何形状在相机空间中对物体进行完整的3D重建。包括第二台相机可以解决剩余的歧义。使用合成数据对该方法进行定量评估并使用真实数据进行了测试,并显示了众所周知的LineMod数据集的其他结果。在机器人实验中,对象的成功掌握证明了其在现实环境中的可用性,并在可能的情况下提供了与其他方法的比较。该方法适用于3D对象模型(例如CAD模型或点云)可用并获得2D图像的精确像素分割图。与其他方法不同,该方法不使用3D深度进行训练,从而扩大了应用领域。
We propose a method for 3D object reconstruction and 6D-pose estimation from 2D images that uses knowledge about object shape as the primary key. In the proposed pipeline, recognition and labeling of objects in 2D images deliver 2D segment silhouettes that are compared with the 2D silhouettes of projections obtained from various views of a 3D model representing the recognized object class. By computing transformation parameters directly from the 2D images, the number of free parameters required during the registration process is reduced, making the approach feasible. Furthermore, 3D transformations and projective geometry are employed to arrive at a full 3D reconstruction of the object in camera space using a calibrated set up. Inclusion of a second camera allows resolving remaining ambiguities. The method is quantitatively evaluated using synthetic data and tested with real data, and additional results for the well-known Linemod data set are shown. In robot experiments, successful grasping of objects demonstrates its usability in real-world environments, and, where possible, a comparison with other methods is provided. The method is applicable to scenarios where 3D object models, e.g., CAD-models or point clouds, are available and precise pixel-wise segmentation maps of 2D images can be obtained. Different from other methods, the method does not use 3D depth for training, widening the domain of application.