IEEE Trans Image Process. 2023;32:2827-2842. doi: 10.1109/TIP.2023.3274988. Epub 2023 May 22.
Convolutional Neural Networks (CNNs) dominate image processing but suffer from local inductive bias, which is addressed by the transformer framework with its inherent ability to capture global context through self-attention mechanisms. However, how to inherit and integrate their advantages to improve compressed sensing is still an open issue. This paper proposes CSformer, a hybrid framework to explore the representation capacity of local and global features. The proposed approach is well-designed for end-to-end compressive image sensing, composed of adaptive sampling and recovery. In the sampling module, images are measured block-by-block by the learned sampling matrix. In the reconstruction stage, the measurements are projected into an initialization stem, a CNN stem, and a transformer stem. The initialization stem mimics the traditional reconstruction of compressive sensing but generates the initial reconstruction in a learnable and efficient manner. The CNN stem and transformer stem are concurrent, simultaneously calculating fine-grained and long-range features and efficiently aggregating them. Furthermore, we explore a progressive strategy and window-based transformer block to reduce the parameters and computational complexity. The experimental results demonstrate the effectiveness of the dedicated transformer-based architecture for compressive sensing, which achieves superior performance compared to state-of-the-art methods on different datasets. Our codes is available at: https://github.com/Lineves7/CSformer.
卷积神经网络(CNNs)在图像处理中占据主导地位,但存在局部归纳偏差,而这一问题可通过具有捕捉全局上下文的内在能力的转换器框架来解决,这种能力是通过自注意力机制实现的。然而,如何继承和整合它们的优势来改进压缩感知仍然是一个悬而未决的问题。本文提出了 CSformer,这是一种用于探索局部和全局特征表示能力的混合框架。所提出的方法非常适合端到端压缩图像感应,由自适应采样和恢复组成。在采样模块中,图像通过学习的采样矩阵逐块进行测量。在重建阶段,测量值被投影到初始化主干、CNN 主干和转换器主干中。初始化主干模仿传统的压缩感知重建,但以可学习和高效的方式生成初始重建。CNN 主干和转换器主干是并发的,同时计算细粒度和长程特征,并有效地对它们进行聚合。此外,我们探索了一种渐进策略和基于窗口的转换器块,以减少参数和计算复杂度。实验结果表明,专门基于转换器的架构对于压缩感知是有效的,与不同数据集上的最先进方法相比,它具有卓越的性能。我们的代码可在:https://github.com/Lineves7/CSformer 上获得。