IEEE Trans Neural Syst Rehabil Eng. 2024;32:3465-3475. doi: 10.1109/TNSRE.2024.3435460. Epub 2024 Sep 20.
With the recent proliferation of large language models (LLMs), such as Generative Pre-trained Transformers (GPT), there has been a significant shift in exploring human and machine comprehension of semantic language meaning. This shift calls for interdisciplinary research that bridges cognitive science and natural language processing (NLP). This pilot study aims to provide insights into individuals' neural states during a semantic inference reading-comprehension task. We propose jointly analyzing LLMs, eye-gaze, and electroencephalographic (EEG) data to study how the brain processes words with varying degrees of relevance to a keyword during reading. We also use feature engineering to improve the fixation-related EEG data classification while participants read words with high versus low relevance to the keyword. The best validation accuracy in this word-level classification is over 60% across 12 subjects. Words highly relevant to the inference keyword received significantly more eye fixations per word: 1.0584 compared to 0.6576, including words with no fixations. This study represents the first attempt to classify brain states at a word level using LLM-generated labels. It provides valuable insights into human cognitive abilities and Artificial General Intelligence (AGI), and offers guidance for developing potential reading-assisted technologies.
随着大型语言模型(LLM)的大量涌现,如生成式预训练转换器(GPT),人们对人类和机器对语义语言意义的理解的探索发生了重大转变。这种转变需要将认知科学和自然语言处理(NLP)相结合的跨学科研究。这项初步研究旨在深入了解个体在语义推理阅读理解任务中的神经状态。我们建议联合分析 LLM、眼动和脑电图(EEG)数据,以研究大脑在阅读过程中如何处理与关键词相关程度不同的单词。我们还使用特征工程来改进与注视相关的 EEG 数据分类,同时参与者阅读与关键词高度相关和低度相关的单词。在 12 名受试者中,该单词级分类的最佳验证准确率超过 60%。与推理关键词高度相关的单词每个单词接收的注视次数明显更多:1.0584 次与 0.6576 次相比,包括没有注视的单词。这项研究首次尝试使用 LLM 生成的标签对大脑状态进行分类。它为人类认知能力和人工智能(AGI)提供了有价值的见解,并为开发潜在的阅读辅助技术提供了指导。