Institute Bioinformatics and Medical Informatics (IBMI), University of Tübingen, Germany; Bernstein Center for Computational Neuroscience, University of Tübingen, Germany; Center for Neuroscience and Artificial Intelligence, BCM, Houston, TX, USA.
Department of Neuroscience, Baylor College of Medicine, Houston, TX, USA; Center for Neuroscience and Artificial Intelligence, BCM, Houston, TX, USA; Department of Electrical and Computer Engineering, Rice University, Houston, TX, USA.
Neuron. 2019 Sep 25;103(6):967-979. doi: 10.1016/j.neuron.2019.08.034.
Despite enormous progress in machine learning, artificial neural networks still lag behind brains in their ability to generalize to new situations. Given identical training data, differences in generalization are caused by many defining features of a learning algorithm, such as network architecture and learning rule. Their joint effect, called "inductive bias," determines how well any learning algorithm-or brain-generalizes: robust generalization needs good inductive biases. Artificial networks use rather nonspecific biases and often latch onto patterns that are only informative about the statistics of the training data but may not generalize to different scenarios. Brains, on the other hand, generalize across comparatively drastic changes in the sensory input all the time. We highlight some shortcomings of state-of-the-art learning algorithms compared to biological brains and discuss several ideas about how neuroscience can guide the quest for better inductive biases by providing useful constraints on representations and network architecture.
尽管机器学习取得了巨大进展,但在泛化新情况的能力方面,人工神经网络仍然落后于大脑。在给定相同的训练数据的情况下,泛化的差异是由学习算法的许多定义特征引起的,例如网络架构和学习规则。它们的共同作用称为“归纳偏差”,决定了任何学习算法或大脑的泛化能力:稳健的泛化需要良好的归纳偏差。人工网络使用的偏差相当不具体,并且经常会抓住仅与训练数据的统计信息有关但可能无法推广到不同场景的模式。另一方面,大脑一直在对感觉输入的相对较大变化进行泛化。我们强调了最先进的学习算法与生物大脑相比存在的一些缺点,并讨论了一些关于神经科学如何通过对表示和网络架构提供有用的约束来指导寻找更好的归纳偏差的想法。