Parr Thomas, Pezzulo Giovanni
Wellcome Centre for Human Neuroimaging, Queen Square Institute of Neurology, University College London, London, United Kingdom.
Institute of Cognitive Sciences and Technologies, National Research Council, Rome, Italy.
Front Syst Neurosci. 2021 Nov 5;15:772641. doi: 10.3389/fnsys.2021.772641. eCollection 2021.
While machine learning techniques have been transformative in solving a range of problems, an important challenge is to understand why they arrive at the decisions they output. Some have argued that this necessitates augmenting machine intelligence with understanding such that, when queried, a machine is able to explain its behaviour (i.e., explainable AI). In this article, we address the issue of machine understanding from the perspective of active inference. This paradigm enables decision making based upon a model of how data are generated. The generative model contains those variables required to explain sensory data, and its inversion may be seen as an attempt to explain the causes of these data. Here we are interested in explanations of one's own actions. This implies a deep generative model that includes a model of the world, used to infer policies, and a higher-level model that attempts to predict which policies will be selected based upon a space of hypothetical (i.e., counterfactual) explanations-and which can subsequently be used to provide (retrospective) explanations about the policies pursued. We illustrate the construct validity of this notion of understanding in relation to human understanding by highlighting the similarities in computational architecture and the consequences of its dysfunction.
虽然机器学习技术在解决一系列问题方面带来了变革,但一个重要的挑战是理解它们为何做出其输出的决策。一些人认为,这需要用理解来增强机器智能,以便在被询问时,机器能够解释其行为(即可解释人工智能)。在本文中,我们从主动推理的角度探讨机器理解问题。这种范式能够基于数据生成方式的模型进行决策。生成模型包含解释感官数据所需的那些变量,其反转可被视为解释这些数据成因的一种尝试。这里我们感兴趣的是对自身行为的解释。这意味着一个深度生成模型,它包括一个用于推断策略的世界模型,以及一个更高层次的模型,该模型试图根据假设(即反事实)解释的空间预测将选择哪些策略——随后可用于提供关于所采用策略的(回顾性)解释。我们通过强调计算架构的相似性及其功能障碍的后果,说明了这种理解概念与人类理解相关的建构效度。