• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于土耳其语放射学报告中命名实体识别的深度学习

Deep learning for named entity recognition in Turkish radiology reports.

作者信息

Abdullahi Abubakar Ahmad, Ganiz Murat Can, Koç Ural, Gökhan Muhammet Batuhan, Aydın Ceren, Özdemir Ali Bahadır

机构信息

Marmara University Faculty of Engineering, Department of Computer Engineering, İstanbul, Türkiye.

Ankara Bilkent City Hospital, Clinic of Radiology, Ankara, Türkiye.

出版信息

Diagn Interv Radiol. 2025 Feb 28. doi: 10.4274/dir.2025.243100.

DOI:10.4274/dir.2025.243100
PMID:40018795
Abstract

PURPOSE

The primary objective of this research is to enhance the accuracy and efficiency of information extraction from radiology reports. In addressing this objective, the study aims to develop and evaluate a deep learning framework for named entity recognition (NER).

METHODS

We used a synthetic dataset of 1,056 Turkish radiology reports created and labeled by the radiologists in our research team. Due to privacy concerns, actual patient data could not be used; however, the synthetic reports closely mimic genuine reports in structure and content. We employed the four-stage DYGIE++ model for the experiments. First, we performed token encoding using four bidirectional encoder representations from transformers (BERT) models: BERTurk, BioBERTurk, PubMedBERT, and XLM-RoBERTa. Second, we introduced adaptive span enumeration, considering the word count of a sentence in Turkish. Third, we adopted span graph propagation to generate a multidirectional graph crucial for coreference resolution. Finally, we used a two-layered feed-forward neural network to classify the named entity.

RESULTS

The experiments conducted on the labeled dataset showcase the approach's effectiveness. The study achieved an F1 score of 80.1 for the NER task, with the BioBERTurk model, which is pre-trained on Turkish Wikipedia, radiology reports, and biomedical texts, proving to be the most effective of the four BERT models used in the experiment.

CONCLUSION

We show how different dataset labels affect the model's performance. The results demonstrate the model's ability to handle the intricacies of Turkish radiology reports, providing a detailed analysis of precision, recall, and F1 scores for each label. Additionally, this study compares its findings with related research in other languages.

CLINICAL SIGNIFICANCE

Our approach provides clinicians with more precise and comprehensive insights to improve patient care by extracting relevant information from radiology reports. This innovation in information extraction streamlines the diagnostic process and helps expedite patient treatment decisions.

摘要

目的

本研究的主要目标是提高从放射学报告中提取信息的准确性和效率。为实现这一目标,该研究旨在开发并评估一种用于命名实体识别(NER)的深度学习框架。

方法

我们使用了由我们研究团队的放射科医生创建并标注的1056份土耳其语放射学报告的合成数据集。出于隐私考虑,无法使用实际患者数据;然而,合成报告在结构和内容上紧密模仿真实报告。我们采用四阶段的DYGIE++模型进行实验。首先,我们使用来自四种基于变换器的双向编码器表征(BERT)模型进行词元编码:BERTurk、BioBERTurk、PubMedBERT和XLM-RoBERTa。其次,考虑到土耳其语句子的单词数量,我们引入了自适应跨度枚举。第三,我们采用跨度图传播来生成对共指消解至关重要的多向图。最后,我们使用两层前馈神经网络对命名实体进行分类。

结果

在标注数据集上进行的实验展示了该方法的有效性。该研究在NER任务中取得了80.1的F1分数,其中在土耳其语维基百科、放射学报告和生物医学文本上预训练的BioBERTurk模型被证明是实验中使用的四个BERT模型中最有效的。

结论

我们展示了不同的数据集标签如何影响模型的性能。结果证明了该模型处理土耳其语放射学报告复杂性的能力,为每个标签提供了精确率、召回率和F1分数的详细分析。此外,本研究将其结果与其他语言的相关研究进行了比较。

临床意义

我们的方法通过从放射学报告中提取相关信息,为临床医生提供更精确和全面的见解,以改善患者护理。这种信息提取方面的创新简化了诊断过程,并有助于加快患者治疗决策。

相似文献

1
Deep learning for named entity recognition in Turkish radiology reports.用于土耳其语放射学报告中命名实体识别的深度学习
Diagn Interv Radiol. 2025 Feb 28. doi: 10.4274/dir.2025.243100.
2
Prescription of Controlled Substances: Benefits and Risks管制药品的处方:益处与风险
3
Knowledge Graph-Enhanced Deep Learning Model (H-SYSTEM) for Hypertensive Intracerebral Hemorrhage: Model Development and Validation.用于高血压性脑出血的知识图谱增强深度学习模型(H-SYSTEM):模型开发与验证
J Med Internet Res. 2025 Jun 12;27:e66055. doi: 10.2196/66055.
4
From BERT to generative AI - Comparing encoder-only vs. large language models in a cohort of lung cancer patients for named entity recognition in unstructured medical reports.从BERT到生成式人工智能——在一组肺癌患者中比较仅编码器模型与大语言模型用于非结构化医疗报告中的命名实体识别
Comput Biol Med. 2025 Sep;195:110665. doi: 10.1016/j.compbiomed.2025.110665. Epub 2025 Jun 24.
5
Radiology report generation using automatic keyword adaptation, frequency-based multi-label classification and text-to-text large language models.使用自动关键词适配、基于频率的多标签分类和文本到文本的大语言模型生成放射学报告。
Comput Biol Med. 2025 Jul 3;196(Pt A):110625. doi: 10.1016/j.compbiomed.2025.110625.
6
Cognitive decline assessment using semantic linguistic content and transformer deep learning architecture.使用语义语言内容和变压器深度学习架构评估认知能力下降。
Int J Lang Commun Disord. 2024 May-Jun;59(3):1110-1127. doi: 10.1111/1460-6984.12973. Epub 2023 Nov 16.
7
Development of a Large-Scale Dataset of Chest Computed Tomography Reports in Japanese and a High-Performance Finding Classification Model: Dataset Development and Validation Study.日语胸部计算机断层扫描报告大规模数据集的开发及高性能发现分类模型:数据集开发与验证研究
JMIR Med Inform. 2025 Aug 28;13:e71137. doi: 10.2196/71137.
8
A fake news detection model using the integration of multimodal attention mechanism and residual convolutional network.一种融合多模态注意力机制和残差卷积网络的假新闻检测模型。
Sci Rep. 2025 Jul 1;15(1):20544. doi: 10.1038/s41598-025-05702-w.
9
Enhancing Pulmonary Disease Prediction Using Large Language Models With Feature Summarization and Hybrid Retrieval-Augmented Generation: Multicenter Methodological Study Based on Radiology Report.使用具有特征总结和混合检索增强生成功能的大语言模型增强肺部疾病预测:基于放射学报告的多中心方法学研究
J Med Internet Res. 2025 Jun 11;27:e72638. doi: 10.2196/72638.
10
Developing an ICD-10 Coding Assistant: Pilot Study Using RoBERTa and GPT-4 for Term Extraction and Description-Based Code Selection.开发国际疾病分类第十版(ICD - 10)编码助手:使用RoBERTa和GPT - 4进行术语提取和基于描述的代码选择的试点研究
JMIR Form Res. 2025 Feb 11;9:e60095. doi: 10.2196/60095.