Fuzzy Logic Systems Institute, Iizuka, Fukuoka 820-0067, Japan.
Neural Netw. 2018 Jan;97:152-161. doi: 10.1016/j.neunet.2017.10.005. Epub 2017 Nov 7.
The neocognitron is a deep (multi-layered) convolutional neural network that can be trained to recognize visual patterns robustly. In the intermediate layers of the neocognitron, local features are extracted from input patterns. In the deepest layer, based on the features extracted in the intermediate layers, input patterns are classified into classes. A method called IntVec (interpolating-vector) is used for this purpose. This paper proposes a new learning rule called margined Winner-Take-All (mWTA) for training the deepest layer. Every time when a training pattern is presented during the learning, if the result of recognition by WTA (Winner-Take-All) is an error, a new cell is generated in the deepest layer. Here we put a certain amount of margin to the WTA. In other words, only during the learning, a certain amount of handicap is given to cells of classes other than that of the training vector, and the winner is chosen under this handicap. By introducing the margin to the WTA, we can generate a compact set of cells, with which a high recognition rate can be obtained with a small computational cost. The ability of this mWTA is demonstrated by computer simulation.
神经认知机是一种深度(多层)卷积神经网络,它可以被训练来稳健地识别视觉模式。在神经认知机的中间层,从输入模式中提取局部特征。在最深层,根据中间层提取的特征,将输入模式分类为不同的类别。为此,使用了一种称为 IntVec(插值向量)的方法。本文提出了一种新的学习规则,称为 margined Winner-Take-All(mWTA),用于训练最深层。每次在学习过程中呈现一个训练模式时,如果 WTA(Winner-Take-All)的识别结果是错误的,那么在最深层会生成一个新的细胞。这里我们给 WTA 加上一定的裕量。换句话说,只有在学习过程中,才会给训练向量以外的类别的细胞一定的 handicap,然后在这个 handicap 下选择 winner。通过在 WTA 中引入裕量,我们可以生成一组紧凑的细胞,用它们可以以较小的计算成本获得较高的识别率。通过计算机模拟证明了这种 mWTA 的能力。