Department of Electrical and Systems Engineering, Washington University in St. Louis, St. Louis, MO, 63130, U.S.A.
Neural Comput. 2019 May;31(5):943-979. doi: 10.1162/neco_a_01182. Epub 2019 Mar 18.
A key aspect of the neural coding problem is understanding how representations of afferent stimuli are built through the dynamics of learning and adaptation within neural networks. The infomax paradigm is built on the premise that such learning attempts to maximize the mutual information between input stimuli and neural activities. In this letter, we tackle the problem of such information-based neural coding with an eye toward two conceptual hurdles. Specifically, we examine and then show how this form of coding can be achieved with online input processing. Our framework thus obviates the biological incompatibility of optimization methods that rely on global network awareness and batch processing of sensory signals. Central to our result is the use of variational bounds as a surrogate objective function, an established technique that has not previously been shown to yield online policies. We obtain learning dynamics for both linear-continuous and discrete spiking neural encoding models under the umbrella of linear gaussian decoders. This result is enabled by approximating certain information quantities in terms of neuronal activity via pairwise feedback mechanisms. Furthermore, we tackle the problem of how such learning dynamics can be realized with strict energetic constraints. We show that endowing networks with auxiliary variables that evolve on a slower timescale can allow for the realization of saddle-point optimization within the neural dynamics, leading to neural codes with favorable properties in terms of both information and energy.
神经编码问题的一个关键方面是理解在神经网络的学习和适应动态中,如何通过输入刺激的动态构建代表。信息最大化范式建立在这样的前提之上,即这种学习试图最大化输入刺激和神经活动之间的互信息。在这封信中,我们着眼于两个概念上的障碍来解决基于信息的神经编码问题。具体来说,我们研究并展示了这种编码形式如何通过在线输入处理来实现。我们的框架因此避免了依赖于全局网络意识和对感觉信号进行批处理的优化方法的生物学不兼容性。我们的结果的核心是使用变分界限作为替代目标函数,这是一种以前没有被证明可以产生在线策略的已有技术。我们在线性高斯解码器的框架下,为线性连续和离散尖峰神经编码模型获得了学习动力学。这一结果是通过通过成对反馈机制用神经元活动来近似某些信息量来实现的。此外,我们还解决了如何在严格的能量约束下实现这种学习动力学的问题。我们表明,为网络赋予在较慢时间尺度上演变的辅助变量,可以在神经动力学中实现鞍点优化,从而在信息和能量方面都具有有利性质的神经编码。