Chen Guijun, Zhang Xueying, Zhang Jing, Li Fenglian, Duan Shufei
College of Information and Computer, Taiyuan University of Technology, Taiyuan, China.
Front Neurorobot. 2022 Sep 30;16:995552. doi: 10.3389/fnbot.2022.995552. eCollection 2022.
Brain-computer interface (BCI) can translate intentions directly into instructions and greatly improve the interaction experience for disabled people or some specific interactive applications. To improve the efficiency of BCI, the objective of this study is to explore the feasibility of an audio-assisted visual BCI speller and a deep learning-based single-trial event related potentials (ERP) decoding strategy.
In this study, a two-stage BCI speller combining the motion-onset visual evoked potential (mVEP) and semantically congruent audio evoked ERP was designed to output the target characters. In the first stage, the different group of characters were presented in the different locations of visual field simultaneously and the stimuli were coded to the mVEP based on a new space division multiple access scheme. And then, the target character can be output based on the audio-assisted mVEP in the second stage. Meanwhile, a spatial-temporal attention-based convolutional neural network (STA-CNN) was proposed to recognize the single-trial ERP components. The CNN can learn 2-dimentional features including the spatial information of different activated channels and time dependence among ERP components. In addition, the STA mechanism can enhance the discriminative event-related features by adaptively learning probability weights.
The performance of the proposed two-stage audio-assisted visual BCI paradigm and STA-CNN model was evaluated using the Electroencephalogram (EEG) recorded from 10 subjects. The average classification accuracy of proposed STA-CNN can reach 59.6 and 77.7% for the first and second stages, which were always significantly higher than those of the comparison methods ( < 0.05).
The proposed two-stage audio-assisted visual paradigm showed a great potential to be used to BCI speller. Moreover, through the analysis of the attention weights from time sequence and spatial topographies, it was proved that STA-CNN could effectively extract interpretable spatiotemporal EEG features.
脑机接口(BCI)能够将意图直接转化为指令,极大地改善残疾人或某些特定交互应用的交互体验。为提高BCI的效率,本研究的目的是探索一种音频辅助视觉BCI拼写器和基于深度学习的单次试验事件相关电位(ERP)解码策略的可行性。
在本研究中,设计了一种结合运动起始视觉诱发电位(mVEP)和语义一致音频诱发ERP的两阶段BCI拼写器来输出目标字符。在第一阶段,不同组的字符同时呈现在视野的不同位置,并且基于一种新的空分多址方案将刺激编码为mVEP。然后,在第二阶段基于音频辅助的mVEP输出目标字符。同时,提出了一种基于时空注意力的卷积神经网络(STA-CNN)来识别单次试验ERP成分。该卷积神经网络可以学习二维特征,包括不同激活通道的空间信息以及ERP成分之间的时间依赖性。此外,STA机制可以通过自适应学习概率权重来增强与事件相关的判别特征。
使用从10名受试者记录的脑电图(EEG)对所提出的两阶段音频辅助视觉BCI范式和STA-CNN模型的性能进行了评估。所提出的STA-CNN在第一阶段和第二阶段的平均分类准确率分别可达59.6%和77.7%,始终显著高于比较方法(<0.05)。
所提出的两阶段音频辅助视觉范式显示出用于BCI拼写器的巨大潜力。此外,通过对时间序列和空间地形图的注意力权重分析,证明了STA-CNN能够有效地提取可解释的时空脑电特征。