School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201600, China.
Faculty of Information Technology, HUTECH University, Ho Chi Minh City, Vietnam.
Sensors (Basel). 2022 Dec 9;22(24):9647. doi: 10.3390/s22249647.
Robust and accurate visual feature tracking is essential for good pose estimation in visual odometry. However, in fast-moving scenes, feature point extraction and matching are unstable because of blurred images and large image disparity. In this paper, we propose an unsupervised monocular visual odometry framework based on a fusion of features extracted from two sources, that is, the optical flow network and the traditional point feature extractor. In the training process, point features are generated for scene images and the outliers of matched point pairs are filtered by FlannMatch. Meanwhile, the optical flow network constrained by the principle of forward-backward flow consistency is used to select another group of corresponding point pairs. The Euclidean distance between the matching points found by FlannMatch and the corresponding point pairs by the flow network is added to the loss function of the flow network. Compared with SURF, the trained flow network shows more robust performance in complicated fast-motion scenarios. Furthermore, we propose the AvgFlow estimation module, which selects one group of the matched point pairs generated by the two methods according to the scene motion. The camera pose is then recovered by Perspective-n-Point (PnP) or the epipolar geometry. Experiments conducted on the KITTI Odometry dataset verify the effectiveness of the trajectory estimation of our approach, especially in fast-moving scenarios.
稳健、准确的视觉特征跟踪对于视觉里程计中的良好姿态估计至关重要。然而,在快速移动的场景中,由于图像模糊和大的图像视差,特征点提取和匹配不稳定。在本文中,我们提出了一种基于融合两种来源的特征的无监督单目视觉里程计框架,即光流网络和传统的点特征提取器。在训练过程中,为场景图像生成点特征,并通过 FlannMatch 过滤匹配点对中的异常值。同时,利用前向-后向流一致性原则约束的光流网络选择另一组对应点对。将 FlannMatch 找到的匹配点与流网络的对应点对之间的欧几里得距离添加到流网络的损失函数中。与 SURF 相比,经过训练的流网络在复杂的快速运动场景中表现出更稳健的性能。此外,我们提出了AvgFlow 估计模块,根据场景运动从两种方法生成的匹配点对中选择一组。然后通过透视 n 点(PnP)或极线几何恢复相机姿态。在 KITTI 里程计数据集上进行的实验验证了我们方法的轨迹估计的有效性,特别是在快速移动的场景中。