Cao Jinming, Li Yangyan, Sun Mingchao, Chen Ying, Lischinski Dani, Cohen-Or Daniel, Chen Baoquan, Tu Changhe
IEEE Trans Image Process. 2022;31:3726-3736. doi: 10.1109/TIP.2022.3175432. Epub 2022 May 26.
Convolutional layers are the core building blocks of Convolutional Neural Networks (CNNs). In this paper, we propose to augment a convolutional layer with an additional depthwise convolution, where each input channel is convolved with a different 2D kernel. The composition of the two convolutions constitutes an over-parameterization, since it adds learnable parameters, while the resulting linear operation can be expressed by a single convolution layer. We refer to this depthwise over-parameterized convolutional layer as DO-Conv, which is a novel way of over-parameterization. We show with extensive experiments that the mere replacement of conventional convolutional layers with DO-Conv layers boosts the performance of CNNs on many classical vision tasks, such as image classification, detection, and segmentation. Moreover, in the inference phase, the depthwise convolution is folded into the conventional convolution, reducing the computation to be exactly equivalent to that of a convolutional layer without over-parameterization. As DO-Conv introduces performance gains without incurring any computational complexity increase for inference, we advocate it as an alternative to the conventional convolutional layer. We open sourced an implementation of DO-Conv in Tensorflow, PyTorch and GluonCV at https://github.com/yangyanli/DO-Conv.
卷积层是卷积神经网络(CNN)的核心构建模块。在本文中,我们建议用额外的深度卷积来增强卷积层,其中每个输入通道都与一个不同的二维内核进行卷积。这两个卷积的组合构成了一种过参数化,因为它增加了可学习参数,而得到的线性运算可以由单个卷积层表示。我们将这种深度过参数化卷积层称为DO-Conv,这是一种新颖的过参数化方式。我们通过大量实验表明,仅用DO-Conv层替换传统卷积层就能提高CNN在许多经典视觉任务上的性能,如图像分类、检测和分割。此外,在推理阶段,深度卷积被折叠到传统卷积中,将计算量减少到与没有过参数化的卷积层完全相同。由于DO-Conv在不增加推理计算复杂度的情况下提高了性能,我们提倡将其作为传统卷积层的替代方案。我们在https://github.com/yangyanli/DO-Conv上开源了DO-Conv在Tensorflow、PyTorch和GluonCV中的实现。