Suppr超能文献

3D2SeqViews:通过具有层次注意力聚合的卷积神经网络聚合顺序视图以进行3D全局特征学习

3D2SeqViews: Aggregating Sequential Views for 3D Global Feature Learning by CNN With Hierarchical Attention Aggregation.

作者信息

Han Zhizhong, Lu Honglei, Liu Zhenbao, Vong Chi-Man, Liu Yu-Shen, Zwicker Matthias, Han Junwei, Chen C L Philip

出版信息

IEEE Trans Image Process. 2019 Aug;28(8):3986-3999. doi: 10.1109/TIP.2019.2904460. Epub 2019 Mar 12.

Abstract

Learning 3D global features by aggregating multiple views is important. Pooling is widely used to aggregate views in deep learning models. However, pooling disregards a lot of content information within views and the spatial relationship among the views, which limits the discriminability of learned features. To resolve this issue, 3D to Sequential Views (3D2SeqViews) is proposed to more effectively aggregate the sequential views using convolutional neural networks with a novel hierarchical attention aggregation. Specifically, the content information within each view is first encoded. Then, the encoded view content information and the sequential spatiality among the views are simultaneously aggregated by the hierarchical attention aggregation, where view-level attention and class-level attention are proposed to hierarchically weight sequential views and shape classes. View-level attention is learned to indicate how much attention is paid to each view by each shape class, which subsequently weights sequential views through a novel recursive view integration. Recursive view integration learns the semantic meaning of view sequence, which is robust to the first view position. Furthermore, class-level attention is introduced to describe how much attention is paid to each shape class, which innovatively employs the discriminative ability of the fine-tuned network. 3D2SeqViews learns more discriminative features than the state-of-the-art, which leads to the outperforming results in shape classification and retrieval under three large-scale benchmarks.

摘要

通过聚合多个视图来学习3D全局特征很重要。池化在深度学习模型中被广泛用于聚合视图。然而,池化忽略了视图中的大量内容信息以及视图之间的空间关系,这限制了所学习特征的可辨别性。为了解决这个问题,提出了3D到序列视图(3D2SeqViews),以使用具有新颖分层注意力聚合的卷积神经网络更有效地聚合序列视图。具体来说,首先对每个视图中的内容信息进行编码。然后,通过分层注意力聚合同时聚合编码后的视图内容信息和视图之间的序列空间性,其中提出了视图级注意力和类级注意力来分层加权序列视图并塑造类别。学习视图级注意力以指示每个形状类别对每个视图的关注程度,随后通过新颖的递归视图整合对序列视图进行加权。递归视图整合学习视图序列的语义含义,这对第一个视图位置具有鲁棒性。此外,引入类级注意力来描述对每个形状类别的关注程度,它创新性地利用了微调网络的辨别能力。3D2SeqViews比现有技术学习到更具辨别力的特征,这在三个大规模基准测试下的形状分类和检索中带来了优于其他方法的结果。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验