Department of Computer Science, University of Maryland, College Park, MD, USA.
Department of Elec. Engr. and Comp. Sci., Syracuse University, Syracuse, NY, USA.
Neural Netw. 2021 Jun;138:78-97. doi: 10.1016/j.neunet.2021.01.031. Epub 2021 Feb 11.
Compositionality refers to the ability of an intelligent system to construct models out of reusable parts. This is critical for the productivity and generalization of human reasoning, and is considered a necessary ingredient for human-level artificial intelligence. While traditional symbolic methods have proven effective for modeling compositionality, artificial neural networks struggle to learn systematic rules for encoding generalizable structured models. We suggest that this is due in part to short-term memory that is based on persistent maintenance of activity patterns without fast weight changes. We present a recurrent neural network that encodes structured representations as systems of contextually-gated dynamical attractors called attractor graphs. This network implements a functionally compositional working memory that is manipulated using top-down gating and fast local learning. We evaluate this approach with empirical experiments on storage and retrieval of graph-based data structures, as well as an automated hierarchical planning task. Our results demonstrate that compositional structures can be stored in and retrieved from neural working memory without persistent maintenance of multiple activity patterns. Further, memory capacity is improved by the use of a fast store-erase learning rule that permits controlled erasure and mutation of previously learned associations. We conclude that the combination of top-down gating and fast associative learning provides recurrent neural networks with a robust functional mechanism for compositional working memory.
组合性是指智能系统能够构建可重用部分的模型的能力。这对于人类推理的生产力和泛化至关重要,并且被认为是人类水平人工智能的必要组成部分。虽然传统的符号方法已被证明对于建模组合性是有效的,但人工神经网络难以学习编码可泛化的结构化模型的系统规则。我们认为,这部分是由于基于持久维持活动模式而没有快速权重变化的短期记忆。我们提出了一种递归神经网络,该网络将结构化表示编码为称为吸引子图的上下文门控动态吸引子系统。该网络实现了使用自上而下的门控和快速局部学习来操纵的功能组合工作记忆。我们通过对基于图的数据结构的存储和检索以及自动分层规划任务的经验实验来评估这种方法。我们的结果表明,组合结构可以存储在神经工作记忆中并从中检索,而无需持久维持多个活动模式。此外,通过使用允许受控擦除和突变先前学习的关联的快速存储-擦除学习规则,提高了存储容量。我们得出结论,自上而下的门控和快速联想学习的结合为递归神经网络提供了用于组合工作记忆的稳健功能机制。