Chen Shunxing, Xiao Guobao, Guo Junwen, Wu Qiangqiang, Ma Jiayi
IEEE Trans Image Process. 2024;33:6002-6015. doi: 10.1109/TIP.2024.3477916. Epub 2024 Oct 22.
We present a novel deep hypergraph modeling architecture (called DHM-Net) for feature matching in this paper. Our network focuses on learning reliable correspondences between two sets of initial feature points by establishing a dynamic hypergraph structure that models group-wise relationships and assigns weights to each node. Compared to existing feature matching methods that only consider pair-wise relationships via a simple graph, our dynamic hypergraph is capable of modeling nonlinear higher-order group-wise relationships among correspondences in an interaction capturing and attention representation learning fashion. Specifically, we propose a novel Deep Hypergraph Modeling block, which initializes an overall hypergraph by utilizing neighbor information, and then adopts node-to-hyperedge and hyperedge-to-node strategies to propagate interaction information among correspondences while assigning weights based on hypergraph attention. In addition, we propose a Differentiation Correspondence-Aware Attention mechanism to optimize the hypergraph for promoting representation learning. The proposed mechanism is able to effectively locate the exact position of the object of importance via the correspondence aware encoding and simple feature gating mechanism to distinguish candidates of inliers. In short, we learn such a dynamic hypergraph format that embeds deep group-wise interactions to explicitly infer categories of correspondences. To demonstrate the effectiveness of DHM-Net, we perform extensive experiments on both real-world outdoor and indoor datasets. Particularly, experimental results show that DHM-Net surpasses the state-of-the-art method by a sizable margin. Our approach obtains an 11.65% improvement under error threshold of 5° for relative pose estimation task on YFCC100M dataset. Code will be released at https://github.com/CSX777/DHM-Net.
在本文中,我们提出了一种用于特征匹配的新型深度超图建模架构(称为DHM-Net)。我们的网络专注于通过建立一个动态超图结构来学习两组初始特征点之间的可靠对应关系,该结构对组内关系进行建模并为每个节点分配权重。与现有的仅通过简单图考虑成对关系的特征匹配方法相比,我们的动态超图能够以交互捕获和注意力表示学习的方式对对应关系中的非线性高阶组内关系进行建模。具体来说,我们提出了一种新颖的深度超图建模模块,它通过利用邻居信息初始化一个整体超图,然后采用节点到超边和超边到节点的策略在对应关系之间传播交互信息,同时基于超图注意力分配权重。此外,我们提出了一种差异化对应感知注意力机制来优化超图以促进表示学习。所提出的机制能够通过对应感知编码和简单特征门控机制有效地定位重要对象的确切位置,以区分内点候选。简而言之,我们学习这样一种动态超图格式,它嵌入深度组内交互以明确推断对应关系的类别。为了证明DHM-Net的有效性,我们在真实世界的室外和室内数据集上进行了广泛的实验。特别是,实验结果表明DHM-Net以相当大的优势超过了现有方法。在YFCC100M数据集上的相对姿态估计任务中,我们的方法在5°误差阈值下获得了11.65%的改进。代码将在https://github.com/CSX /777/DHM-Net上发布。