National Scientific and Technical Research Council (CONICET), Godoy Cruz 2290, CABA 1425, Argentina; Institute of Applied and Interdisciplinary Physics and Department of Physics, University of Buenos Aires, Pabellón 1, Ciudad Universitaria, CABA 1425, Argentina.
National Scientific and Technical Research Council (CONICET), Godoy Cruz 2290, CABA 1425, Argentina; Departamento de Computación, Facultad de Ciencias Exactas y Naturales, Universidad de Buenos Aires, Pabellón 1, Ciudad Universitaria, CABA 1425, Argentina; Instituto de Investigación en Ciencias de la Computación (ICC), CONICET-Universidad de Buenos Aires, Pabellón 1, Ciudad Universitaria, CABA 1425, Argentina.
Neuroimage. 2024 Jun;293:120626. doi: 10.1016/j.neuroimage.2024.120626. Epub 2024 Apr 25.
Spatio-temporal patterns of evoked brain activity contain information that can be used to decode and categorize the semantic content of visual stimuli. However, this procedure can be biased by low-level image features independently of the semantic content present in the stimuli, prompting the need to understand the robustness of different models regarding these confounding factors. In this study, we trained machine learning models to distinguish between concepts included in the publicly available THINGS-EEG dataset using electroencephalography (EEG) data acquired during a rapid serial visual presentation paradigm. We investigated the contribution of low-level image features to decoding accuracy in a multivariate model, utilizing broadband data from all EEG channels. Additionally, we explored a univariate model obtained through data-driven feature selection applied to the spatial and frequency domains. While the univariate models exhibited better decoding accuracy, their predictions were less robust to the confounding effect of low-level image statistics. Notably, some of the models maintained their accuracy even after random replacement of the training dataset with semantically unrelated samples that presented similar low-level content. In conclusion, our findings suggest that model optimization impacts sensitivity to confounding factors, regardless of the resulting classification performance. Therefore, the choice of EEG features for semantic decoding should ideally be informed by criteria beyond classifier performance, such as the neurobiological mechanisms under study.
诱发脑活动的时空模式包含可用于解码和分类视觉刺激语义内容的信息。然而,这种方法可能会受到与刺激中呈现的语义内容无关的低水平图像特征的影响,因此需要了解不同模型对这些混杂因素的稳健性。在这项研究中,我们使用脑电图(EEG)数据,通过快速序列视觉呈现范式,对公开可用的 THINGS-EEG 数据集进行训练,利用机器学习模型来区分概念。我们研究了低水平图像特征在多元模型中对解码精度的贡献,利用来自所有 EEG 通道的宽带数据。此外,我们还探索了通过应用于空间和频率域的数据驱动特征选择获得的单变量模型。虽然单变量模型表现出更好的解码精度,但它们的预测对低水平图像统计数据的混杂效应的鲁棒性较差。值得注意的是,一些模型即使在使用语义上不相关的样本随机替换训练数据集后,仍能保持准确性,这些样本呈现出相似的低水平内容。总之,我们的研究结果表明,无论分类性能如何,模型优化都会影响对混杂因素的敏感性。因此,用于语义解码的 EEG 特征的选择应理想地基于分类器性能以外的标准,例如正在研究的神经生物学机制。