Sun Xiaoxia, Nasrabadi Nasser M, Tran Trac D
IEEE Trans Image Process. 2019 Jul 17. doi: 10.1109/TIP.2019.2928121.
In this paper, we propose a novel deep sparse coding network (SCN) capable of efficiently adapting its own regularization parameters for a given application. The network is trained end-to-end with a supervised task-driven learning algorithm via error backpropagation. During training, the network learns both the dictionaries and the regularization parameters of each sparse coding layer so that the reconstructive dictionaries are smoothly transformed into increasingly discriminative representations. In addition, the adaptive regularization also offers the network more flexibility to adjust sparsity levels. Furthermore, we have devised a sparse coding layer utilizing a 'skinny' dictionary. Integral to computational efficiency, these skinny dictionaries compress the high dimensional sparse codes into lower dimensional structures. The adaptivity and discriminability of our fifteen-layer sparse coding network are demonstrated on five benchmark datasets, namely Cifar-10, Cifar-100, STL-10, SVHN and MNIST, most of which are considered difficult for sparse coding models. Experimental results show that our architecture overwhelmingly outperforms traditional one-layer sparse coding architectures while using much fewer parameters. Moreover, our multilayer architecture exploits the benefits of depth with sparse coding's characteristic ability to operate on smaller datasets. In such data-constrained scenarios, our technique demonstrates highly competitive performance compared to the deep neural networks.
在本文中,我们提出了一种新颖的深度稀疏编码网络(SCN),它能够针对给定应用有效地自适应调整自身的正则化参数。该网络通过误差反向传播,使用监督任务驱动的学习算法进行端到端训练。在训练过程中,网络学习每个稀疏编码层的字典和正则化参数,以便将重构字典平滑地转换为更具判别力的表示。此外,自适应正则化还为网络提供了更大的灵活性来调整稀疏度水平。此外,我们设计了一种使用“瘦”字典的稀疏编码层。这些瘦字典对于计算效率至关重要,它们将高维稀疏码压缩为低维结构。我们的十五层稀疏编码网络的适应性和判别力在五个基准数据集上得到了验证,即Cifar-10、Cifar-100、STL-10、SVHN和MNIST,其中大多数数据集对于稀疏编码模型来说都被认为具有挑战性。实验结果表明,我们的架构在使用少得多的参数的情况下,远远优于传统的单层稀疏编码架构。此外,我们的多层架构利用了深度的优势以及稀疏编码在较小数据集上运行的独特能力。在这种数据受限的场景中,与深度神经网络相比,我们的技术展现出极具竞争力的性能。