Wang Jiake, Guan Yong, Kang Zhenjia, Chen Pengzhan
School of Intelligent Manufacturing, Taizhou University, Taizhou 318000, China.
School of Electrical and Automation Engineering, East China Jiaotong University, Nanchang 330013, China.
Sensors (Basel). 2024 Jun 27;24(13):4178. doi: 10.3390/s24134178.
Visual ranging technology holds great promise in various fields such as unmanned driving and robot navigation. However, complex dynamic environments pose significant challenges to its accuracy and robustness. Existing monocular visual ranging methods are susceptible to scale uncertainty, while binocular visual ranging is sensitive to changes in lighting and texture. To overcome the limitations of single visual ranging, this paper proposes a fusion method for monocular and binocular visual ranging based on an adaptive Unscented Kalman Filter (AUKF). The proposed method first utilizes a monocular camera to estimate the initial distance based on the pixel size, and then employs the triangulation principle with a binocular camera to obtain accurate depth. Building upon this foundation, a probabilistic fusion framework is constructed to dynamically fuse monocular and binocular ranging using the AUKF. The AUKF employs nonlinear recursive filtering to estimate the optimal distance and its uncertainty, and introduces an adaptive noise-adjustment mechanism to dynamically update the observation noise based on fusion residuals, thus suppressing outlier interference. Additionally, an adaptive fusion strategy based on depth hypothesis propagation is designed to autonomously adjust the noise prior of the AUKF by combining current environmental features and historical measurement information, further enhancing the algorithm's adaptability to complex scenes. To validate the effectiveness of the proposed method, comprehensive evaluations were conducted on large-scale public datasets such as KITTI and complex scene data collected in real-world scenarios. The quantitative results demonstrate that the fusion method significantly improves the overall accuracy and stability of visual ranging, reducing the average relative error within an 8 m range by 43.1% and 40.9% compared to monocular and binocular ranging, respectively. Compared to traditional methods, the proposed method significantly enhances ranging accuracy and exhibits stronger robustness against factors such as lighting changes and dynamic targets. The sensitivity analysis further confirmed the effectiveness of the AUKF framework and adaptive noise strategy. In summary, the proposed fusion method effectively combines the advantages of monocular and binocular vision, significantly expanding the application range of visual ranging technology in intelligent driving, robotics, and other fields while ensuring accuracy, robustness, and real-time performance.
视觉测距技术在无人驾驶和机器人导航等各个领域都具有巨大的潜力。然而,复杂的动态环境对其准确性和鲁棒性构成了重大挑战。现有的单目视觉测距方法容易受到尺度不确定性的影响,而双目视觉测距则对光照和纹理的变化敏感。为了克服单一视觉测距的局限性,本文提出了一种基于自适应无迹卡尔曼滤波器(AUKF)的单目和双目视觉测距融合方法。该方法首先利用单目相机根据像素大小估计初始距离,然后利用双目相机的三角测量原理获得精确深度。在此基础上,构建了一个概率融合框架,使用AUKF动态融合单目和双目测距。AUKF采用非线性递归滤波来估计最优距离及其不确定性,并引入自适应噪声调整机制,根据融合残差动态更新观测噪声,从而抑制异常值干扰。此外,设计了一种基于深度假设传播的自适应融合策略,通过结合当前环境特征和历史测量信息自主调整AUKF的噪声先验,进一步提高算法对复杂场景的适应性。为了验证所提方法的有效性,在大规模公共数据集如KITTI以及实际场景中收集的复杂场景数据上进行了综合评估。定量结果表明,该融合方法显著提高了视觉测距的整体准确性和稳定性,与单目和双目测距相比,在8米范围内平均相对误差分别降低了43.1%和40.9%。与传统方法相比,所提方法显著提高了测距精度,并且对光照变化和动态目标等因素表现出更强的鲁棒性。敏感性分析进一步证实了AUKF框架和自适应噪声策略的有效性。综上所述,所提融合方法有效地结合了单目和双目视觉的优点,在确保准确性、鲁棒性和实时性能的同时,显著扩展了视觉测距技术在智能驾驶、机器人等领域的应用范围。