• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

意义的几何学:评估基于不同Transformer模型的句子嵌入用于自然语言推理

The geometry of meaning: evaluating sentence embeddings from diverse transformer-based models for natural language inference.

作者信息

Alsuhaibani Mohammed

机构信息

Department of Computer Science, College of Computer, Qassim University, Buraydah, Saudi Arabia.

出版信息

PeerJ Comput Sci. 2025 Jun 16;11:e2957. doi: 10.7717/peerj-cs.2957. eCollection 2025.

DOI:10.7717/peerj-cs.2957
PMID:40567742
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12193426/
Abstract

Natural language inference (NLI) is a fundamental task in natural language processing that focuses on determining the relationship between pairs of sentences. In this article, we present a simple and straightforward approach to evaluate the effectiveness of various transformer-based models such as bidirectional encoder representations from transformers (BERT), Generative Pre-trained Transformer (GPT), robustly optimized BERT approach (RoBERTa), and XLNet in generating sentence embeddings for NLI. We conduct comprehensive experiments with different pooling techniques and evaluate the embeddings using different norms across multiple layers of each model. Our results demonstrate that the choice of pooling strategy, norm, and model layer significantly impacts the performance of NLI, with the best results achieved using max pooling and the L2 norm across specific model layers. On the Stanford Natural Language Inference (SNLI) dataset, the model reached 90% accuracy and 86% F1-score, while on the MedNLI dataset, the highest F1-score recorded was 84%. This article provides insights into how different models and evaluation strategies can be effectively combined to improve the understanding and classification of sentence relationships in NLI tasks.

摘要

自然语言推理(NLI)是自然语言处理中的一项基础任务,专注于确定句子对之间的关系。在本文中,我们提出一种简单直接的方法,来评估各种基于Transformer的模型的有效性,例如来自Transformer的双向编码器表示(BERT)、生成式预训练Transformer(GPT)、稳健优化的BERT方法(RoBERTa)以及XLNet在为自然语言推理生成句子嵌入方面的有效性。我们使用不同的池化技术进行了全面的实验,并在每个模型的多个层中使用不同的范数来评估嵌入。我们的结果表明,池化策略、范数和模型层的选择对自然语言推理的性能有显著影响,在特定模型层使用最大池化和L2范数可取得最佳结果。在斯坦福自然语言推理(SNLI)数据集上,该模型达到了90%的准确率和86%的F1分数,而在医学自然语言推理(MedNLI)数据集上,记录的最高F1分数为84%。本文深入探讨了如何有效地结合不同的模型和评估策略,以提高自然语言推理任务中句子关系的理解和分类。

相似文献

1
The geometry of meaning: evaluating sentence embeddings from diverse transformer-based models for natural language inference.意义的几何学:评估基于不同Transformer模型的句子嵌入用于自然语言推理
PeerJ Comput Sci. 2025 Jun 16;11:e2957. doi: 10.7717/peerj-cs.2957. eCollection 2025.
2
Detecting Redundant Health Survey Questions by Using Language-Agnostic Bidirectional Encoder Representations From Transformers Sentence Embedding: Algorithm Development Study.使用来自Transformer句子嵌入的语言无关双向编码器表示法检测冗余健康调查问题:算法开发研究
JMIR Med Inform. 2025 Jun 10;13:e71687. doi: 10.2196/71687.
3
Sentiment Analysis Using a Large Language Model-Based Approach to Detect Opioids Mixed With Other Substances Via Social Media: Method Development and Validation.使用基于大语言模型的方法通过社交媒体检测与其他物质混合的阿片类药物的情感分析:方法开发与验证
JMIR Infodemiology. 2025 Jun 19;5:e70525. doi: 10.2196/70525.
4
Large Language Model Architectures in Health Care: Scoping Review of Research Perspectives.医疗保健中的大语言模型架构:研究视角的范围综述
J Med Internet Res. 2025 Jun 19;27:e70315. doi: 10.2196/70315.
5
Trajectory-Ordered Objectives for Self-Supervised Representation Learning of Temporal Healthcare Data Using Transformers: Model Development and Evaluation Study.使用Transformer进行时间序列医疗数据自监督表示学习的轨迹有序目标:模型开发与评估研究
JMIR Med Inform. 2025 Jun 4;13:e68138. doi: 10.2196/68138.
6
Enhancing Pulmonary Disease Prediction Using Large Language Models With Feature Summarization and Hybrid Retrieval-Augmented Generation: Multicenter Methodological Study Based on Radiology Report.使用具有特征总结和混合检索增强生成功能的大语言模型增强肺部疾病预测:基于放射学报告的多中心方法学研究
J Med Internet Res. 2025 Jun 11;27:e72638. doi: 10.2196/72638.
7
Signs and symptoms to determine if a patient presenting in primary care or hospital outpatient settings has COVID-19.在基层医疗机构或医院门诊环境中,如果患者出现以下症状和体征,可判断其是否患有 COVID-19。
Cochrane Database Syst Rev. 2022 May 20;5(5):CD013665. doi: 10.1002/14651858.CD013665.pub3.
8
The potential of Generative Pre-trained Transformer 4 (GPT-4) to analyse medical notes in three different languages: a retrospective model-evaluation study.生成式预训练变换器4(GPT-4)分析三种不同语言医学笔记的潜力:一项回顾性模型评估研究。
Lancet Digit Health. 2025 Jan;7(1):e35-e43. doi: 10.1016/S2589-7500(24)00246-2.
9
Text intelligent correction in English translation: A study on integrating models with dependency attention mechanism.英文翻译中的文本智能校正:一项关于集成具有依存注意力机制模型的研究。
PLoS One. 2025 Jun 24;20(6):e0319690. doi: 10.1371/journal.pone.0319690. eCollection 2025.
10
Native language identification from text using a fine-tuned GPT-2 model.使用微调的GPT-2模型从文本中识别母语。
PeerJ Comput Sci. 2025 May 28;11:e2909. doi: 10.7717/peerj-cs.2909. eCollection 2025.

本文引用的文献

1
AmericasNLI: Machine translation and natural language inference systems for Indigenous languages of the Americas.美洲自然语言推理项目:用于美洲原住民语言的机器翻译和自然语言推理系统。
Front Artif Intell. 2022 Dec 2;5:995667. doi: 10.3389/frai.2022.995667. eCollection 2022.
2
Legal Information Retrieval and Entailment Based on BM25, Transformer and Semantic Thesaurus Methods.基于BM25、Transformer和语义词库方法的法律信息检索与蕴含
Rev Socionetwork Strateg. 2022;16(1):157-174. doi: 10.1007/s12626-022-00103-1. Epub 2022 Feb 7.
3
Jointly learning word embeddings using a corpus and a knowledge base.
联合使用语料库和知识库学习词向量。
PLoS One. 2018 Mar 12;13(3):e0193094. doi: 10.1371/journal.pone.0193094. eCollection 2018.