Center for Computational Neuroscience and Neural Technology, Boston University, Boston, MA 02215, USA.
Neural Comput. 2013 May;25(5):1261-76. doi: 10.1162/NECO_a_00435. Epub 2013 Mar 7.
Convolutional models of object recognition achieve invariance to spatial transformations largely because of the use of a suitably defined pooling operator. This operator typically takes the form of a max or average function defined across units tuned to the same feature. As a model of the brain's ventral pathway, where computations are carried out by weighted synaptic connections, such pooling can lead to spatial invariance only if the weights that connect similarly tuned units to a given pooling unit are of approximately equal strengths. How identical weights can be learned in the face of nonuniformly distributed data remains unclear. In this letter, we show how various versions of the trace learning rule can help solve this problem. This allows us in turn to explain previously published results and make recommendations as to the optimal rule for invariance learning.
对象识别的卷积模型在很大程度上实现了对空间变换的不变性,这主要是因为使用了适当定义的池化操作符。该操作符通常采用在对同一特征调谐的单元上定义的最大值或平均值函数的形式。作为大脑腹侧通路的模型,其中计算是通过加权突触连接进行的,这种池化只能导致空间不变性,前提是连接到给定池化单元的类似调谐单元的权重具有大致相等的强度。在面对非均匀分布的数据时,如何学习相同的权重仍然不清楚。在这封信中,我们展示了轨迹学习规则的各种版本如何帮助解决这个问题。这反过来又使我们能够解释以前发表的结果,并就不变性学习的最佳规则提出建议。