Shin Unsub, Yi Eunkyung, Song Sanghoun
Department of Linguistics, Korea University, Seoul, Republic of Korea.
Department of English Education, Ewha Womans University, Seoul, Republic of Korea.
Front Psychol. 2023 Feb 23;14:937656. doi: 10.3389/fpsyg.2023.937656. eCollection 2023.
The recent success of deep learning neural language models such as Bidirectional Encoder Representations from Transformers (BERT) has brought innovations to computational language research. The present study explores the possibility of using a language model in investigating human language processes, based on the case study of negative polarity items (NPIs). We first conducted an experiment with BERT to examine whether the model successfully captures the hierarchical structural relationship between an NPI and its licensor and whether it may lead to an error analogous to the grammatical illusion shown in the psycholinguistic experiment (Experiment 1). We also investigated whether the language model can capture the fine-grained semantic properties of NPI licensors and discriminate their subtle differences on the scale of licensing strengths (Experiment 2). The results of the two experiments suggest that overall, the neural language model is highly sensitive to both syntactic and semantic constraints in NPI processing. The model's processing patterns and sensitivities are shown to be very close to humans, suggesting their role as a research tool or object in the study of language.
诸如来自变换器的双向编码器表示(BERT)等深度学习神经语言模型最近的成功给计算语言研究带来了创新。本研究基于对负极性项(NPI)的案例研究,探讨了使用语言模型来研究人类语言过程的可能性。我们首先用BERT进行了一项实验,以检验该模型是否成功捕捉到了一个负极性项与其许可语之间的层次结构关系,以及它是否可能导致类似于心理语言学实验中所显示的语法错觉的错误(实验1)。我们还研究了语言模型是否能够捕捉负极性项许可语的细粒度语义属性,并在许可强度的尺度上区分它们的细微差异(实验2)。这两个实验的结果表明,总体而言,神经语言模型在负极性项处理中对句法和语义约束都高度敏感。该模型的处理模式和敏感度显示出与人类非常接近,表明了它们在语言研究中作为研究工具或对象的作用。