Lakshmanan Karthik C, Sadtler Patrick T, Tyler-Kabara Elizabeth C, Batista Aaron P, Yu Byron M
Robotics Institute and Center for the Neural Basis of Cognition, Carnegie Mellon University, Pittsburgh, PA 15213, U.S.A.
Department of Bioengineering, Center for the Neural Basis of Cognition, and Systems Neuroscience Institute, University of Pittsburgh, Pittsburgh, PA 15261, U.S.A.
Neural Comput. 2015 Sep;27(9):1825-56. doi: 10.1162/NECO_a_00759. Epub 2015 Jun 16.
Noisy, high-dimensional time series observations can often be described by a set of low-dimensional latent variables. Commonly used methods to extract these latent variables typically assume instantaneous relationships between the latent and observed variables. In many physical systems, changes in the latent variables manifest as changes in the observed variables after time delays. Techniques that do not account for these delays can recover a larger number of latent variables than are present in the system, thereby making the latent representation more difficult to interpret. In this work, we introduce a novel probabilistic technique, time-delay gaussian-process factor analysis (TD-GPFA), that performs dimensionality reduction in the presence of a different time delay between each pair of latent and observed variables. We demonstrate how using a gaussian process to model the evolution of each latent variable allows us to tractably learn these delays over a continuous domain. Additionally, we show how TD-GPFA combines temporal smoothing and dimensionality reduction into a common probabilistic framework. We present an expectation/conditional maximization either (ECME) algorithm to learn the model parameters. Our simulations demonstrate that when time delays are present, TD-GPFA is able to correctly identify these delays and recover the latent space. We then applied TD-GPFA to the activity of tens of neurons recorded simultaneously in the macaque motor cortex during a reaching task. TD-GPFA is able to better describe the neural activity using a more parsimonious latent space than GPFA, a method that has been used to interpret motor cortex data but does not account for time delays. More broadly, TD-GPFA can help to unravel the mechanisms underlying high-dimensional time series data by taking into account physical delays in the system.
嘈杂的高维时间序列观测值通常可以用一组低维潜在变量来描述。提取这些潜在变量的常用方法通常假设潜在变量和观测变量之间存在瞬时关系。在许多物理系统中,潜在变量的变化会在经过时间延迟后表现为观测变量的变化。不考虑这些延迟的技术可能会恢复出比系统中实际存在的更多的潜在变量,从而使潜在表示更难解释。在这项工作中,我们引入了一种新颖的概率技术——时延高斯过程因子分析(TD-GPFA),它在每对潜在变量和观测变量之间存在不同时间延迟的情况下进行降维。我们展示了如何使用高斯过程对每个潜在变量的演化进行建模,从而使我们能够在连续域上易于处理地学习这些延迟。此外,我们展示了TD-GPFA如何将时间平滑和降维结合到一个通用的概率框架中。我们提出了一种期望/条件最大化(ECME)算法来学习模型参数。我们的模拟表明,当存在时间延迟时,TD-GPFA能够正确识别这些延迟并恢复潜在空间。然后,我们将TD-GPFA应用于猕猴在伸手任务期间同时记录的数十个神经元的活动。与GPFA相比,TD-GPFA能够使用更简洁的潜在空间更好地描述神经活动,GPFA是一种用于解释运动皮层数据但不考虑时间延迟的方法。更广泛地说,TD-GPFA可以通过考虑系统中的物理延迟来帮助揭示高维时间序列数据背后的机制。