ATR Brain Information Communication Research Group, Computational Neuroscience Laboratory, Hikaridai, Kyoto, 619-0288, Japan.
Biol Cybern. 2021 Oct;115(5):415-430. doi: 10.1007/s00422-021-00904-7.
In several papers published in Biological Cybernetics in the 1980s and 1990s, Kawato and colleagues proposed computational models explaining how internal models are acquired in the cerebellum. These models were later supported by neurophysiological experiments using monkeys and neuroimaging experiments involving humans. These early studies influenced neuroscience from basic, sensory-motor control to higher cognitive functions. One of the most perplexing enigmas related to internal models is to understand the neural mechanisms that enable animals to learn large-dimensional problems with so few trials. Consciousness and metacognition-the ability to monitor one's own thoughts, may be part of the solution to this enigma. Based on literature reviews of the past 20 years, here we propose a computational neuroscience model of metacognition. The model comprises a modular hierarchical reinforcement-learning architecture of parallel and layered, generative-inverse model pairs. In the prefrontal cortex, a distributed executive network called the "cognitive reality monitoring network" (CRMN) orchestrates conscious involvement of generative-inverse model pairs in perception and action. Based on mismatches between computations by generative and inverse models, as well as reward prediction errors, CRMN computes a "responsibility signal" that gates selection and learning of pairs in perception, action, and reinforcement learning. A high responsibility signal is given to the pairs that best capture the external world, that are competent in movements (small mismatch), and that are capable of reinforcement learning (small reward-prediction error). CRMN selects pairs with higher responsibility signals as objects of metacognition, and consciousness is determined by the entropy of responsibility signals across all pairs. This model could lead to new-generation AI, which exhibits metacognition, consciousness, dimension reduction, selection of modules and corresponding representations, and learning from small samples. It may also lead to the development of a new scientific paradigm that enables the causal study of consciousness by combining CRMN and decoded neurofeedback.
在 20 世纪 80 年代和 90 年代发表在《生物控制论》上的几篇论文中,川田和他的同事们提出了计算模型,解释了内部模型如何在小脑获得。这些模型后来得到了使用猴子的神经生理学实验和涉及人类的神经影像学实验的支持。这些早期的研究从基础的感觉运动控制到更高的认知功能,影响了神经科学。与内部模型相关的最令人费解的谜团之一是理解动物如何在如此少的试次内学习大维问题的神经机制。意识和元认知——监测自己思维的能力,可能是解决这个谜团的一部分。基于过去 20 年的文献综述,我们在这里提出了一个元认知的计算神经科学模型。该模型由一个模块化的分层强化学习架构组成,包括并行和分层的生成-逆模型对。在前额叶皮层中,一个称为“认知现实监测网络”(CRMN)的分布式执行网络协调生成-逆模型对在感知和行动中的有意识参与。基于生成模型和逆模型计算之间的差异以及奖励预测误差,CRMN 计算出一个“责任信号”,该信号门控对感知、行动和强化学习中模型对的选择和学习。高责任信号赋予最能捕捉外部世界、动作能力强(小的不匹配)且能够强化学习(小的奖励预测误差)的模型对。CRMN 选择具有更高责任信号的模型对作为元认知的对象,意识由所有模型对的责任信号熵决定。该模型可能会导致新一代人工智能的出现,这种人工智能表现出元认知、意识、维度降低、模块和相应表示的选择,以及从小样本中学习。它还可能导致一种新的科学范式的发展,该范式通过结合 CRMN 和解码神经反馈来实现对意识的因果研究。