Suppr超能文献

基于双注意力和多损失的遥感影像耕地侵占检测与建筑物提取

Cropland encroachment detection dual attention and multi-loss based building extraction in remote sensing images.

作者信息

Wang Junshu, Cai Mingrui, Gu Yifan, Liu Zhen, Li Xiaoxin, Han Yuxing

机构信息

College of Electronic Engineering, College of Artificial Intelligence, South China Agricultural University, Guangzhou, China.

Shenzhen International Graduate School, Tsinghua University, Shenzhen, China.

出版信息

Front Plant Sci. 2022 Sep 6;13:993961. doi: 10.3389/fpls.2022.993961. eCollection 2022.

Abstract

The United Nations predicts that by 2050, the world's total population will increase to 9.15 billion, but the per capita cropland will drop to 0.151°hm. The acceleration of urbanization often comes at the expense of the encroachment of cropland, the unplanned expansion of urban area has adversely affected cultivation. Therefore, the automatic extraction of buildings, which are the main carriers of urban population activities, in remote sensing images has become a more meaningful cropland observation task. To solve the shortcomings of traditional building extraction methods such as insufficient utilization of image information, relying on manual characterization, etc. A U-Net based deep learning building extraction model is proposed and named AttsegGAN. This study proposes an adversarial loss based on the Generative Adversarial Network in terms of training strategy, and the additionally trained learnable discriminator is used as a distance measurer for the two probability distributions of ground truth and prediction . In addition, for the sharpness of the building edge, the Sobel edge loss based on the Sobel operator is weighted and jointly participated in the training. In WHU building dataset, this study applies the components and strategies step by step, and verifies their effectiveness. Furthermore, the addition of the attention module is also subjected to ablation experiments and the final framework is determined. Compared with the original, AttsegGAN improved by 0.0062, 0.0027, and 0.0055 on Acc, F1, and IoU respectively after adopting all improvements. In the comparative experiment. AttsegGAN is compared with state-of-the-arts including U-Net, DeeplabV3+, PSPNet, and DANet on both WHU and Massachusetts building dataset. In WHU dataset, AttsegGAN achieved 0.9875, 0.9435, and 0.8907 on Acc, F1, and IoU, surpassed U-Net by 0.0260, 0.1183, and 0.1883, respectively, demonstrated the effectiveness of the proposed components in a similar hourglass structure. In Massachusetts dataset, AttsegGAN also surpassed state-of-the-arts, achieved 0.9395, 0.8328, and 0.7130 on Acc, F1, and IoU, respectively, it improved IoU by 0.0412 over the second-ranked PSPNet, and it was 0.0025 and 0.0101 higher than the second place in Acc and F1.

摘要

联合国预测,到2050年,世界总人口将增至91.5亿,但人均耕地将降至0.151公顷。城市化加速往往以侵占耕地为代价,城市区域的无序扩张对耕种产生了不利影响。因此,在遥感影像中自动提取作为城市人口活动主要载体的建筑物,已成为一项更具意义的耕地观测任务。为解决传统建筑物提取方法存在的图像信息利用不充分、依赖人工特征描述等缺点,提出了一种基于U-Net的深度学习建筑物提取模型并命名为AttsegGAN。本研究在训练策略方面提出了基于生成对抗网络的对抗损失,将额外训练的可学习判别器用作真实值和预测值两种概率分布的距离测量器。此外,针对建筑物边缘的清晰度,对基于Sobel算子的Sobel边缘损失进行加权并共同参与训练。在WHU建筑物数据集中,本研究逐步应用各组件和策略,并验证了其有效性。此外,还对注意力模块的添加进行了消融实验并确定了最终框架。在采用所有改进措施后,与原始模型相比,AttsegGAN在Acc、F1和IoU上分别提高了0.0062、0.0027和0.0055。在对比实验中,将AttsegGAN与包括U-Net、DeeplabV3+、PSPNet和DANet在内的当前最优模型在WHU和马萨诸塞州建筑物数据集上进行了比较。在WHU数据集中,AttsegGAN在Acc、F1和IoU上分别达到了0.9875、0.9435和0.8907,分别比U-Net高出0.0260、0.1183和0.1883,证明了所提出组件在类似沙漏结构中的有效性。在马萨诸塞州数据集中,AttsegGAN也超越了当前最优模型,在Acc、F1和IoU上分别达到了0.9395、0.8328和0.7130,其IoU比排名第二的PSPNet提高了0.0412,在Acc和F1上比第二名分别高出0.0025和0.0101。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec53/9486080/5f7b935d8719/fpls-13-993961-g001.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验