论文标题

颗粒FM:利用用于将移动相机定位在野外的密集点轨迹

ParticleSfM: Exploiting Dense Point Trajectories for Localizing Moving Cameras in the Wild

论文作者

Zhao, Wang, Liu, Shaohui, Guo, Hengkai, Wang, Wenping, Liu, Yong-Jin

论文摘要

从单眼视频中估算移动摄像头的姿势是一个具有挑战性的问题,尤其是由于动态环境中移动对象的存在,在动态环境中,现有摄像头姿势估计方法的性能易于几何一致的像素。为了应对这一挑战,我们为视频提供了一种强大的密集间接结构,该结构是基于由成对光流初始化的密集对应的。我们的关键想法是将远程视频对应性优化为密集的点轨迹,并使用它来了解运动分割的强大估计。提出了一种新型的神经网络结构来处理不规则的点轨迹数据。然后,通过将远程点轨迹分类为静态的远程轨迹的部分进行估算和优化的摄像头姿势。 MPI Sintel数据集的实验表明,与现有最新方法相比,我们的系统产生的相机轨迹明显更准确。此外,我们的方法能够在完全静态的场景上保留相机姿势的合理准确性,该场景始终超过了具有端到端深度学习的基于最先进的密度对应方法,这证明了基于光流和点轨迹的密集间接方法的潜力。由于点轨迹表示是一般的,我们进一步提出了具有动态对象的复杂运动的野外单眼视频的结果和比较。代码可在https://github.com/bytedance/particle-sfm上找到。

Estimating the pose of a moving camera from monocular video is a challenging problem, especially due to the presence of moving objects in dynamic environments, where the performance of existing camera pose estimation methods are susceptible to pixels that are not geometrically consistent. To tackle this challenge, we present a robust dense indirect structure-from-motion method for videos that is based on dense correspondence initialized from pairwise optical flow. Our key idea is to optimize long-range video correspondence as dense point trajectories and use it to learn robust estimation of motion segmentation. A novel neural network architecture is proposed for processing irregular point trajectory data. Camera poses are then estimated and optimized with global bundle adjustment over the portion of long-range point trajectories that are classified as static. Experiments on MPI Sintel dataset show that our system produces significantly more accurate camera trajectories compared to existing state-of-the-art methods. In addition, our method is able to retain reasonable accuracy of camera poses on fully static scenes, which consistently outperforms strong state-of-the-art dense correspondence based methods with end-to-end deep learning, demonstrating the potential of dense indirect methods based on optical flow and point trajectories. As the point trajectory representation is general, we further present results and comparisons on in-the-wild monocular videos with complex motion of dynamic objects. Code is available at https://github.com/bytedance/particle-sfm.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源