• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

沙普利习语:分析用于通用习语词元识别的BERT句子嵌入

Shapley Idioms: Analysing BERT Sentence Embeddings for General Idiom Token Identification.

作者信息

Nedumpozhimana Vasudevan, Klubička Filip, Kelleher John D

机构信息

ADAPT Centre, Technological University Dublin, Dublin, Ireland.

出版信息

Front Artif Intell. 2022 Mar 14;5:813967. doi: 10.3389/frai.2022.813967. eCollection 2022.

DOI:10.3389/frai.2022.813967
PMID:35360661
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8964145/
Abstract

This article examines the basis of Natural Language Understanding of transformer based language models, such as BERT. It does this through a case study on idiom token classification. We use idiom token identification as a basis for our analysis because of the variety of information types that have previously been explored in the literature for this task, including: topic, lexical, and syntactic features. This variety of relevant information types means that the task of idiom token identification enables us to explore the forms of linguistic information that a BERT language model captures and encodes in its representations. The core of this article presents three experiments. The first experiment analyzes the effectiveness of BERT sentence embeddings for creating a general idiom token identification model and the results indicate that the BERT sentence embeddings outperform Skip-Thought. In the second and third experiment we use the game theory concept of Shapley Values to rank the usefulness of individual idiomatic expressions for model training and use this ranking to analyse the type of information that the model finds useful. We find that a combination of idiom-intrinsic and topic-based properties contribute to an expression's usefulness in idiom token identification. Overall our results indicate that BERT efficiently encodes a variety of information from topic, through lexical and syntactic information. Based on these results we argue that notwithstanding recent criticisms of language model based semantics, the ability of BERT to efficiently encode a variety of linguistic information types does represent a significant step forward in natural language understanding.

摘要

本文探讨了基于Transformer的语言模型(如BERT)的自然语言理解基础。这是通过一个关于习语词元分类的案例研究来实现的。我们将习语词元识别作为分析的基础,是因为此前文献中针对该任务探讨了多种信息类型,包括:主题、词汇和句法特征。这种多样的相关信息类型意味着习语词元识别任务使我们能够探究BERT语言模型在其表示中捕获和编码的语言信息形式。本文核心部分呈现了三个实验。第一个实验分析了BERT句子嵌入用于创建通用习语词元识别模型的有效性,结果表明BERT句子嵌入优于Skip-Thought。在第二个和第三个实验中,我们使用夏普利值的博弈论概念对各个习语表达在模型训练中的有用性进行排名,并利用该排名来分析模型认为有用的信息类型。我们发现习语内在属性和基于主题的属性相结合有助于一个表达在习语词元识别中的有用性。总体而言,我们的结果表明BERT能有效地从主题开始,通过词汇和句法信息对多种信息进行编码。基于这些结果,我们认为尽管最近对基于语言模型的语义学存在批评,但BERT有效编码多种语言信息类型的能力确实代表了自然语言理解方面的显著进步。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8d8/8964145/ea00d6c01070/frai-05-813967-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8d8/8964145/519e6767b7d0/frai-05-813967-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8d8/8964145/ea00d6c01070/frai-05-813967-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8d8/8964145/519e6767b7d0/frai-05-813967-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8d8/8964145/ea00d6c01070/frai-05-813967-g0002.jpg

相似文献

1
Shapley Idioms: Analysing BERT Sentence Embeddings for General Idiom Token Identification.沙普利习语:分析用于通用习语词元识别的BERT句子嵌入
Front Artif Intell. 2022 Mar 14;5:813967. doi: 10.3389/frai.2022.813967. eCollection 2022.
2
Representing idioms: syntactic and contextual effects on idiom processing.习语的表征:句法和语境对习语加工的影响
Lang Speech. 2013 Sep;56(Pt 3):373-94. doi: 10.1177/0023830913484899.
3
That's the way the cookie bounces: syntactic and semantic components of experimentally elicited idiom blends.事情就是这样:实验诱发的习语混合的句法和语义成分。
Mem Cognit. 1997 Jan;25(1):57-71. doi: 10.3758/bf03197285.
4
Predicting Semantic Similarity Between Clinical Sentence Pairs Using Transformer Models: Evaluation and Representational Analysis.使用Transformer模型预测临床句子对之间的语义相似性:评估与表征分析
JMIR Med Inform. 2021 May 26;9(5):e23099. doi: 10.2196/23099.
5
Improved biomedical word embeddings in the transformer era.Transformer 时代改进的生物医学词向量。
J Biomed Inform. 2021 Aug;120:103867. doi: 10.1016/j.jbi.2021.103867. Epub 2021 Jul 18.
6
An Improved BERT and Syntactic Dependency Representation Model for Sentiment Analysis.基于改进的 BERT 和句法依存关系表示模型的情感分析。
Comput Intell Neurosci. 2022 May 5;2022:5754151. doi: 10.1155/2022/5754151. eCollection 2022.
7
Stacking-BERT model for Chinese medical procedure entity normalization.基于堆叠 BERT 的中文医疗过程实体标准化模型。
Math Biosci Eng. 2023 Jan;20(1):1018-1036. doi: 10.3934/mbe.2023047. Epub 2022 Oct 24.
8
Comprehension of idiomatic expressions: effects of predictability and literality.习语表达的理解:可预测性和字面意义的影响。
J Exp Psychol Learn Mem Cogn. 1994 Sep;20(5):1126-38. doi: 10.1037//0278-7393.20.5.1126.
9
BERT-GT: cross-sentence n-ary relation extraction with BERT and Graph Transformer.BERT-GT:使用BERT和图变换器进行跨句子n元关系提取
Bioinformatics. 2021 Apr 5;36(24):5678-5685. doi: 10.1093/bioinformatics/btaa1087.
10
Idiom comprehension in Alzheimer's disease: the role of the central executive.阿尔茨海默病中的习语理解:中央执行系统的作用。
Brain. 2003 Nov;126(Pt 11):2419-30. doi: 10.1093/brain/awg243. Epub 2003 Aug 5.