Cui Yiming, Zhang Wei-Nan, Che Wanxiang, Liu Ting, Chen Zhigang, Wang Shijin
Research Center for Social Computing and Information Retrieval, Harbin Institute of Technology, Harbin 150001, China.
State Key Laboratory of Cognitive Intelligence, iFLYTEK Research, Beijing 100083, China.
iScience. 2022 Mar 31;25(5):104176. doi: 10.1016/j.isci.2022.104176. eCollection 2022 May 20.
Achieving human-level performance on some of the machine reading comprehension (MRC) datasets is no longer challenging with the help of powerful pre-trained language models (PLMs). However, the internal mechanism of these artifacts remains unclear, placing an obstacle to further understand these models. This paper focuses on conducting a series of analytical experiments to examine the relations between the multi-head self-attention and the final MRC system performance, revealing the potential explainability in PLM-based MRC models. To ensure the robustness of the analyses, we perform our experiments in a multilingual way on top of various PLMs. We discover that passage-to-question and passage understanding attentions are the most important ones in the question answering process, showing strong correlations to the final performance than other parts. Through comprehensive visualizations and case studies, we also observe several general findings on the attention maps, which can be helpful to understand how these models solve the questions.
借助强大的预训练语言模型(PLM),在一些机器阅读理解(MRC)数据集上实现人类水平的性能已不再具有挑战性。然而,这些模型的内部机制仍不清楚,这为进一步理解这些模型设置了障碍。本文专注于进行一系列分析实验,以检验多头自注意力与最终MRC系统性能之间的关系,揭示基于PLM的MRC模型中的潜在可解释性。为确保分析的稳健性,我们在各种PLM之上以多语言方式进行实验。我们发现,篇章到问题的注意力和篇章理解注意力在问答过程中最为重要,与最终性能的相关性比其他部分更强。通过全面的可视化和案例研究,我们还在注意力图上观察到了几个一般性的发现,这有助于理解这些模型是如何解决问题的。