Politecnico di Torino, Corso Duca degli Abruzzi 24, I-10129 Torino, Italy.
Human Genetics Foundation-Torino, Via Nizza 52, I-10126 Torino, Italy.
Phys Rev Lett. 2015 Sep 18;115(12):128101. doi: 10.1103/PhysRevLett.115.128101.
We show that discrete synaptic weights can be efficiently used for learning in large scale neural systems, and lead to unanticipated computational performance. We focus on the representative case of learning random patterns with binary synapses in single layer networks. The standard statistical analysis shows that this problem is exponentially dominated by isolated solutions that are extremely hard to find algorithmically. Here, we introduce a novel method that allows us to find analytical evidence for the existence of subdominant and extremely dense regions of solutions. Numerical experiments confirm these findings. We also show that the dense regions are surprisingly accessible by simple learning protocols, and that these synaptic configurations are robust to perturbations and generalize better than typical solutions. These outcomes extend to synapses with multiple states and to deeper neural architectures. The large deviation measure also suggests how to design novel algorithmic schemes for optimization based on local entropy maximization.
我们表明,离散的突触权重可以有效地用于大规模神经网络中的学习,并带来意想不到的计算性能。我们专注于用二进制突触在单层网络中学习随机模式的代表性案例。标准的统计分析表明,这个问题受到孤立解的指数控制,这些孤立解极难通过算法找到。在这里,我们引入了一种新的方法,使我们能够找到存在亚主导和极其密集的解区域的分析证据。数值实验证实了这些发现。我们还表明,通过简单的学习协议可以非常容易地到达这些密集区域,并且这些突触配置比典型的解决方案更稳健,泛化效果更好。这些结果扩展到具有多个状态的突触和更深层次的神经架构。大偏差测度也为基于局部熵最大化的优化设计新的算法方案提供了思路。