Das Arun, Mock Jeffrey, Irani Farzan, Huang Yufei, Najafirad Peyman, Golob Edward
Secure AI and Autonomy Laboratory, University of Texas at San Antonio, San Antonio, TX, United States.
UPMC Hillman Cancer Center, University of Pittsburgh Medical Center, Pittsburgh, PA, United States.
Front Neurosci. 2022 Aug 1;16:912798. doi: 10.3389/fnins.2022.912798. eCollection 2022.
A key goal of cognitive neuroscience is to better understand how dynamic brain activity relates to behavior. Such dynamics, in terms of spatial and temporal patterns of brain activity, are directly measured with neurophysiological methods such as EEG, but can also be indirectly expressed by the body. Autonomic nervous system activity is the best-known example, but, muscles in the eyes and face can also index brain activity. Mostly parallel lines of artificial intelligence research show that EEG and facial muscles both encode information about emotion, pain, attention, and social interactions, among other topics. In this study, we examined adults who stutter (AWS) to understand the relations between dynamic brain and facial muscle activity and predictions about future behavior (fluent or stuttered speech). AWS can provide insight into brain-behavior dynamics because they naturally fluctuate between episodes of fluent and stuttered speech behavior. We focused on the period when speech preparation occurs, and used EEG and facial muscle activity measured from video to predict whether the upcoming speech would be fluent or stuttered. An explainable self-supervised multimodal architecture learned the temporal dynamics of both EEG and facial muscle movements during speech preparation in AWS, and predicted fluent or stuttered speech at 80.8% accuracy (chance=50%). Specific EEG and facial muscle signals distinguished fluent and stuttered trials, and systematically varied from early to late speech preparation time periods. The self-supervised architecture successfully identified multimodal activity that predicted upcoming behavior on a trial-by-trial basis. This approach could be applied to understanding the neural mechanisms driving variable behavior and symptoms in a wide range of neurological and psychiatric disorders. The combination of direct measures of neural activity and simple video data may be applied to developing technologies that estimate brain state from subtle bodily signals.
认知神经科学的一个关键目标是更好地理解动态大脑活动与行为之间的关系。就大脑活动的空间和时间模式而言,这种动态可以通过脑电图(EEG)等神经生理学方法直接测量,但也可以由身体间接表现出来。自主神经系统活动是最广为人知的例子,不过,眼睛和面部的肌肉也能够指示大脑活动。大部分人工智能研究的平行方向表明,脑电图和面部肌肉都对情绪、疼痛、注意力以及社会互动等信息进行编码。在本研究中,我们对成年口吃者(AWS)进行了检查,以了解动态大脑活动与面部肌肉活动之间的关系以及对未来行为(流畅或口吃言语)的预测。成年口吃者能够为大脑-行为动态提供见解,因为他们在流畅和口吃言语行为发作之间自然波动。我们关注言语准备阶段,利用从视频中测量得到的脑电图和面部肌肉活动来预测即将到来的言语是流畅还是口吃。一种可解释的自监督多模态架构学习了成年口吃者言语准备期间脑电图和面部肌肉运动的时间动态,并以80.8%的准确率(随机概率为50%)预测了流畅或口吃言语。特定的脑电图和面部肌肉信号区分了流畅和口吃试验,并且在言语准备的早期到晚期时间段内有系统地变化。这种自监督架构成功识别出了逐次预测即将发生行为的多模态活动。这种方法可应用于理解驱动广泛神经和精神疾病中可变行为及症状的神经机制。神经活动的直接测量与简单视频数据的结合可能应用于开发从细微身体信号估计大脑状态的技术。