Kostas Demetres, Aroca-Ouellette Stéphane, Rudzicz Frank
Department Computer Science, University of Toronto, Toronto, ON, Canada.
Vector Institute for Artificial Intelligence, Toronto, ON, Canada.
Front Hum Neurosci. 2021 Jun 23;15:653659. doi: 10.3389/fnhum.2021.653659. eCollection 2021.
Deep neural networks (DNNs) used for brain-computer interface (BCI) classification are commonly expected to learn general features when trained across a variety of contexts, such that these features could be fine-tuned to specific contexts. While some success is found in such an approach, we suggest that this interpretation is limited and an alternative would better leverage the newly (publicly) available massive electroencephalography (EEG) datasets. We consider how to adapt techniques and architectures used for language modeling (LM) that appear capable of ingesting awesome amounts of data toward the development of encephalography modeling with DNNs in the same vein. We specifically adapt an approach effectively used for automatic speech recognition, which similarly (to LMs) uses a self-supervised training objective to learn compressed representations of raw data signals. After adaptation to EEG, we find that a single pre-trained model is capable of modeling completely novel raw EEG sequences recorded with differing hardware, and different subjects performing different tasks. Furthermore, both the internal representations of this model and the entire architecture can be fine-tuned to a of downstream BCI and EEG classification tasks, outperforming prior work in more (sleep stage classification) self-supervision.
用于脑机接口(BCI)分类的深度神经网络(DNN)通常期望在跨多种情境进行训练时学习通用特征,以便这些特征能够针对特定情境进行微调。虽然这种方法取得了一些成功,但我们认为这种解释是有限的,另一种方法可以更好地利用新(公开)可用的大量脑电图(EEG)数据集。我们考虑如何调整用于语言建模(LM)的技术和架构,这些技术和架构似乎能够处理大量数据,并以同样的方式用于开发基于DNN的脑电建模。我们特别采用了一种有效地用于自动语音识别的方法,该方法与语言模型类似,使用自监督训练目标来学习原始数据信号的压缩表示。在将其应用于脑电后,我们发现单个预训练模型能够对使用不同硬件记录的全新原始脑电序列进行建模,以及对执行不同任务的不同受试者进行建模。此外,该模型的内部表示和整个架构都可以针对一系列下游BCI和脑电分类任务进行微调,在更多(睡眠阶段分类)自监督方面优于先前的工作。