Zheng Fuzhong, Wang Xu, Wang Luyao, Zhang Xiong, Zhu Hongze, Wang Long, Zhang Haisu
College of Information and Communication, National University of Defense Technology, Wuhan 430074, China.
Sensors (Basel). 2023 Oct 13;23(20):8437. doi: 10.3390/s23208437.
Due to the swift growth in the scale of remote sensing imagery, scholars have progressively directed their attention towards achieving efficient and adaptable cross-modal retrieval for remote sensing images. They have also steadily tackled the distinctive challenge posed by the multi-scale attributes of these images. However, existing studies primarily concentrate on the characterization of these features, neglecting the comprehensive investigation of the complex relationship between multi-scale targets and the semantic alignment of these targets with text. To address this issue, this study introduces a fine-grained semantic alignment method that adequately aggregates multi-scale information (referred to as FAAMI). The proposed approach comprises multiple stages. Initially, we employ a computing-friendly cross-layer feature connection method to construct a multi-scale feature representation of an image. Subsequently, we devise an efficient feature consistency enhancement module to rectify the incongruous semantic discrimination observed in cross-layer features. Finally, a shallow cross-attention network is employed to capture the fine-grained semantic relationship between multiple-scale image regions and the corresponding words in the text. Extensive experiments were conducted using two datasets: RSICD and RSITMD. The results demonstrate that the performance of FAAMI surpasses that of recently proposed advanced models in the same domain, with significant improvements observed in R@K and other evaluation metrics. Specifically, the mR values achieved by FAAMI are 23.18% and 35.99% for the two datasets, respectively.
由于遥感影像规模的迅速增长,学者们逐渐将注意力转向实现高效且适应性强的遥感影像跨模态检索。他们也在稳步应对这些影像多尺度属性带来的独特挑战。然而,现有研究主要集中于这些特征的表征,而忽视了对多尺度目标之间复杂关系以及这些目标与文本语义对齐的全面研究。为解决这一问题,本研究引入一种能充分聚合多尺度信息的细粒度语义对齐方法(称为FAAMI)。所提出的方法包括多个阶段。首先,我们采用一种计算友好的跨层特征连接方法来构建图像的多尺度特征表示。随后,我们设计了一个高效的特征一致性增强模块来纠正跨层特征中观察到的不一致语义判别。最后,使用一个浅层交叉注意力网络来捕捉多尺度图像区域与文本中相应单词之间的细粒度语义关系。使用两个数据集RSICD和RSITMD进行了广泛的实验。结果表明,FAAMI的性能超过了同一领域最近提出的先进模型,在R@K和其他评估指标上有显著提升。具体而言,FAAMI在两个数据集上分别实现的mR值为23.18%和35.99%。