Xu Jing, Pan Yu, Pan Xinglin, Hoi Steven, Yi Zhang, Xu Zenglin
IEEE Trans Neural Netw Learn Syst. 2023 Nov;34(11):9562-9567. doi: 10.1109/TNNLS.2022.3158966. Epub 2023 Oct 27.
The ResNet and its variants have achieved remarkable successes in various computer vision tasks. Despite its success in making gradient flow through building blocks, the information communication of intermediate layers of blocks is ignored. To address this issue, in this brief, we propose to introduce a regulator module as a memory mechanism to extract complementary features of the intermediate layers, which are further fed to the ResNet. In particular, the regulator module is composed of convolutional recurrent neural networks (RNNs) [e.g., convolutional long short-term memories (LSTMs) or convolutional gated recurrent units (GRUs)], which are shown to be good at extracting spatio-temporal information. We named the new regulated network as regulated residual network (RegNet). The regulator module can be easily implemented and appended to any ResNet architecture. Experimental results on three image classification datasets have demonstrated the promising performance of the proposed architecture compared with the standard ResNet, squeeze-and-excitation ResNet, and other state-of-the-art architectures.
ResNet及其变体在各种计算机视觉任务中取得了显著成功。尽管它在使梯度流经各个模块方面取得了成功,但模块中间层的信息通信却被忽视了。为了解决这个问题,在本简报中,我们建议引入一个调节器模块作为一种记忆机制,以提取中间层的互补特征,并将这些特征进一步输入到ResNet中。具体而言,调节器模块由卷积循环神经网络(RNN)[例如,卷积长短期记忆网络(LSTM)或卷积门控循环单元(GRU)]组成,这些网络在提取时空信息方面表现出色。我们将新的调节网络命名为调节残差网络(RegNet)。调节器模块可以很容易地实现并附加到任何ResNet架构上。在三个图像分类数据集上的实验结果表明,与标准ResNet、挤压激励ResNet和其他现有最先进架构相比,所提出的架构具有良好的性能。