Barto A G
Hum Neurobiol. 1985;4(4):229-56.
Since the usual approaches to cooperative computation in networks of neuron-like computating elements do not assume that network components have any "preferences", they do not make substantive contact with game theoretic concepts, despite their use of some of the same terminology. In the approach presented here, however, each network component, or adaptive element, is a self-interested agent that prefers some inputs over others and "works" toward obtaining the most highly preferred inputs. Here we describe an adaptive element that is robust enough to learn to cooperate with other elements like itself in order to further its self-interests. It is argued that some of the longstanding problems concerning adaptation and learning by networks might be solvable by this form of cooperativity, and computer simulation experiments are described that show how networks of self-interested components that are sufficiently robust can solve rather difficult learning problems. We then place the approach in its proper historical and theoretical perspective through comparison with a number of related algorithms. A secondary aim of this article is to suggest that beyond what is explicitly illustrated here, there is a wealth of ideas from game theory and allied disciplines such as mathematical economics that can be of use in thinking about cooperative computation in both nervous systems and man-made systems.
由于在类神经元计算元件网络中进行协作计算的常用方法并不假定网络组件有任何“偏好”,所以尽管使用了一些相同的术语,但它们与博弈论概念并无实质性关联。然而,在此提出的方法中,每个网络组件或自适应元件都是一个自利的主体,它对某些输入的偏好超过其他输入,并“努力”获取最偏好的输入。这里我们描述一种自适应元件,它足够强大,能够学会与其他类似自身的元件合作以增进自身利益。有人认为,一些关于网络自适应和学习的长期问题可能通过这种合作形式得以解决,并且描述了计算机模拟实验,展示了足够强大的自利组件网络如何能够解决相当困难的学习问题。然后,我们通过与一些相关算法进行比较,将该方法置于恰当的历史和理论视角中。本文的第二个目的是表明,除了这里明确阐述的内容之外,博弈论以及数学经济学等相关学科还有大量思想,可用于思考神经系统和人造系统中的协作计算。