• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

英文翻译中的文本智能校正:一项关于集成具有依存注意力机制模型的研究。

Text intelligent correction in English translation: A study on integrating models with dependency attention mechanism.

作者信息

Liu Yutong, Zhang Shile

机构信息

School of Humanities and Social Sciences, Xi'an Polytechnic University, Xi'an, China.

Shaanxi Contemporary Red Culture Training and Education Center, Xi'an, China.

出版信息

PLoS One. 2025 Jun 24;20(6):e0319690. doi: 10.1371/journal.pone.0319690. eCollection 2025.

DOI:10.1371/journal.pone.0319690
PMID:40554486
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12186988/
Abstract

Improving translation quality and efficiency is one of the key challenges in the field of Natural Language Processing (NLP). This study proposes an enhanced model based on Bidirectional Encoder Representations from Transformers (BERT), combined with a dependency self-attention mechanism, to automatically detect and correct textual errors in the translation process. The model aims to strengthen the understanding of sentence structure, thereby improving both the accuracy and efficiency of error correction. The research uses the Conference on Natural Language Learning (CoNLL)-2014 dataset as an experimental benchmark, which contains a rich collection of grammatical error samples and is a standard resource in linguistic research. During model training, the Adam optimization algorithm is employed, and the model's performance is enhanced by introducing a customized dependency self-attention mechanism for parameter optimization. To validate the model's effectiveness, the performance of the baseline model and the improved model is compared using multiple evaluation metrics, including accuracy, recall, F1 score, edit distance, Bilingual Evaluation Understudy (BLEU) score, and average processing time. The results show that the proposed model significantly outperforms the baseline model in terms of accuracy (improving from 0.78 to 0.85), recall (improving from 0.81 to 0.87), and F1 score (improving from 0.79 to 0.86). The average edit distance decreases from 3.2 to 2.5, the BLEU score increases from 0.65 to 0.72, and the average processing time is reduced from 2.3 seconds to 1.8 seconds. This study provides an innovative approach for intelligent text correction tasks, expands the application scenarios of the BERT model, and offers significant support for the practical implementation of NLP technologies. The findings not only highlight the advantages of the improved model but also offer new ideas and directions for future related research.

摘要

提高翻译质量和效率是自然语言处理(NLP)领域的关键挑战之一。本研究提出了一种基于变换器双向编码器表征(BERT)的增强模型,并结合依存自注意力机制,以在翻译过程中自动检测和纠正文本错误。该模型旨在加强对句子结构的理解,从而提高纠错的准确性和效率。研究使用自然语言学习会议(CoNLL)-2014数据集作为实验基准,该数据集包含丰富的语法错误样本集合,是语言学研究中的标准资源。在模型训练过程中,采用了Adam优化算法,并通过引入定制的依存自注意力机制进行参数优化来提高模型性能。为了验证模型的有效性,使用包括准确率、召回率、F1分数、编辑距离、双语评估替补(BLEU)分数和平均处理时间在内的多个评估指标,比较了基线模型和改进模型的性能。结果表明,所提出的模型在准确率(从0.78提高到0.85)、召回率(从0.81提高到0.87)和F1分数(从0.79提高到0.86)方面显著优于基线模型。平均编辑距离从3.2降至2.5,BLEU分数从0.65提高到0.72,平均处理时间从2.3秒减少到1.8秒。本研究为智能文本纠错任务提供了一种创新方法,扩展了BERT模型的应用场景,并为NLP技术的实际应用提供了重要支持。研究结果不仅突出了改进模型的优势,还为未来相关研究提供了新的思路和方向。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/df69/12186988/46b4403c5755/pone.0319690.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/df69/12186988/a743d0109be1/pone.0319690.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/df69/12186988/c4018c7ccb48/pone.0319690.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/df69/12186988/6bf9f887bf4f/pone.0319690.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/df69/12186988/0e9e31a69c22/pone.0319690.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/df69/12186988/46b4403c5755/pone.0319690.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/df69/12186988/a743d0109be1/pone.0319690.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/df69/12186988/c4018c7ccb48/pone.0319690.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/df69/12186988/6bf9f887bf4f/pone.0319690.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/df69/12186988/0e9e31a69c22/pone.0319690.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/df69/12186988/46b4403c5755/pone.0319690.g005.jpg

相似文献

1
Text intelligent correction in English translation: A study on integrating models with dependency attention mechanism.英文翻译中的文本智能校正:一项关于集成具有依存注意力机制模型的研究。
PLoS One. 2025 Jun 24;20(6):e0319690. doi: 10.1371/journal.pone.0319690. eCollection 2025.
2
Enhancing Pulmonary Disease Prediction Using Large Language Models With Feature Summarization and Hybrid Retrieval-Augmented Generation: Multicenter Methodological Study Based on Radiology Report.使用具有特征总结和混合检索增强生成功能的大语言模型增强肺部疾病预测:基于放射学报告的多中心方法学研究
J Med Internet Res. 2025 Jun 11;27:e72638. doi: 10.2196/72638.
3
Detecting Redundant Health Survey Questions by Using Language-Agnostic Bidirectional Encoder Representations From Transformers Sentence Embedding: Algorithm Development Study.使用来自Transformer句子嵌入的语言无关双向编码器表示法检测冗余健康调查问题:算法开发研究
JMIR Med Inform. 2025 Jun 10;13:e71687. doi: 10.2196/71687.
4
Trajectory-Ordered Objectives for Self-Supervised Representation Learning of Temporal Healthcare Data Using Transformers: Model Development and Evaluation Study.使用Transformer进行时间序列医疗数据自监督表示学习的轨迹有序目标:模型开发与评估研究
JMIR Med Inform. 2025 Jun 4;13:e68138. doi: 10.2196/68138.
5
Edge computing based english translation model using fuzzy semantic optimal control technique.基于模糊语义最优控制技术的边缘计算英语翻译模型。
PLoS One. 2025 Jun 25;20(6):e0320481. doi: 10.1371/journal.pone.0320481. eCollection 2025.
6
Sentiment Analysis Using a Large Language Model-Based Approach to Detect Opioids Mixed With Other Substances Via Social Media: Method Development and Validation.使用基于大语言模型的方法通过社交媒体检测与其他物质混合的阿片类药物的情感分析:方法开发与验证
JMIR Infodemiology. 2025 Jun 19;5:e70525. doi: 10.2196/70525.
7
Adefovir dipivoxil and pegylated interferon alfa-2a for the treatment of chronic hepatitis B: a systematic review and economic evaluation.阿德福韦酯与聚乙二醇化干扰素α-2a治疗慢性乙型肝炎:系统评价与经济学评估
Health Technol Assess. 2006 Aug;10(28):iii-iv, xi-xiv, 1-183. doi: 10.3310/hta10280.
8
Leveraging large language models for spelling correction in Turkish.利用大语言模型进行土耳其语拼写纠正。
PeerJ Comput Sci. 2025 Jun 16;11:e2889. doi: 10.7717/peerj-cs.2889. eCollection 2025.
9
Eliciting adverse effects data from participants in clinical trials.从临床试验参与者中获取不良反应数据。
Cochrane Database Syst Rev. 2018 Jan 16;1(1):MR000039. doi: 10.1002/14651858.MR000039.pub2.
10
Signs and symptoms to determine if a patient presenting in primary care or hospital outpatient settings has COVID-19.在基层医疗机构或医院门诊环境中,如果患者出现以下症状和体征,可判断其是否患有 COVID-19。
Cochrane Database Syst Rev. 2022 May 20;5(5):CD013665. doi: 10.1002/14651858.CD013665.pub3.

本文引用的文献

1
Research on Anomaly Network Detection Based on Self-Attention Mechanism.基于自注意力机制的异常网络检测研究。
Sensors (Basel). 2023 May 25;23(11):5059. doi: 10.3390/s23115059.
2
A Survey of Current Machine Learning Approaches to Student Free-Text Evaluation for Intelligent Tutoring.智能辅导中用于学生自由文本评估的当前机器学习方法综述。
Int J Artif Intell Educ. 2022 Nov 28:1-39. doi: 10.1007/s40593-022-00323-0.
3
Performance of Multiple Pretrained BERT Models to Automate and Accelerate Data Annotation for Large Datasets.多个预训练BERT模型在为大型数据集自动执行和加速数据标注方面的性能。
Radiol Artif Intell. 2022 Jun 29;4(4):e220007. doi: 10.1148/ryai.220007. eCollection 2022 Jul.