Suppr超能文献

Structure-Preserved Self-Attention for Fusion Image Information in Multiple Color Spaces.

作者信息

He Zhu, Lin Mingwei, Luo Xin, Xu Zeshui

出版信息

IEEE Trans Neural Netw Learn Syst. 2025 Jul;36(7):13021-13035. doi: 10.1109/TNNLS.2024.3490800.

Abstract

The selection and utilization of different color spaces significantly impact the recognition performance of deep learning models in downstream tasks. Existing studies typically leverage image information from various color spaces through model integration or channel concatenation. However, these methods result in excessive model size and suboptimal utilization of image information. In this study, we propose the structure-preserved self-attention network (SPSANet) model for efficient fusion of image information from different color spaces. This model incorporates a novel structure-preserved self-attention (SPSA) module that employs a single-head pixel-wise attention mechanism, as opposed to the conventional multihead self-attention (MHSA) approach. Specifically, feature maps from all color space grouping paths are utilized for similarity matching, enabling the model to focus on critical pixel locations across different color spaces. This design mitigates the dependence of the SPSANet model on the choice of color space while enhancing the advantages of integrating multiple color spaces. The SPSANet model also employs channel shuffle operations to facilitate limited interaction between information flows from different color space paths. Experimental results demonstrate that the SPSANet model, utilizing eight common color spaces-RGB, Luv, XYZ, Lab, HSV, YCrCb, YUV, and HLS-achieves superior recognition performance with reduced parameters and computational cost.

摘要

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验