Suppr超能文献

基于多模态信息融合与混合注意力机制的中医命名实体识别

Chinese medical named entity recognition based on multimodal information fusion and hybrid attention mechanism.

作者信息

Luo Zhen, Che Jingping, Ji Fan

机构信息

School of Artificial Intelligence, Zhoukou Normal University, Zhoukou, Henan, China.

出版信息

PLoS One. 2025 Jun 11;20(6):e0325660. doi: 10.1371/journal.pone.0325660. eCollection 2025.

Abstract

Chinese Medical Named Entity Recognition (CMNER) seeks to identify and extract medical entities from unstructured medical texts. Existing methods often depend on single-modality representations and fail to fully exploit the complementary nature of different features. This paper presents a multimodal information fusion-based approach for medical named entity recognition, integrating a hybrid attention mechanism. A Dual-Stream Network architecture is employed to extract multimodal features at both the character and word levels, followed by deep fusion to enhance the model's ability to recognize medical entities. The Cross-Stream Attention mechanism is introduced to facilitate information exchange between different modalities and capture cross-modal global dependencies. Multi-Head Attention is employed to further enhance feature representation and improve the model's ability to delineate medical entity boundaries. The Conditional Random Field (CRF) layer is used for decoding, ensuring global consistency in entity predictions and thereby enhancing recognition accuracy and robustness. The proposed method achieves F1 scores of 65.26%, 80.31%, and 86.73% on the CMeEE-V2, IMCS-V2-NER, and CHIP-STS datasets, respectively, outperforming other models and demonstrating significant improvements in medical entity recognition accuracy and multiple evaluation metrics.

摘要

中文医学命名实体识别(CMNER)旨在从非结构化医学文本中识别和提取医学实体。现有方法通常依赖于单模态表示,无法充分利用不同特征的互补性。本文提出了一种基于多模态信息融合的医学命名实体识别方法,集成了混合注意力机制。采用双流网络架构在字符和单词级别提取多模态特征,然后进行深度融合以增强模型识别医学实体的能力。引入跨流注意力机制以促进不同模态之间的信息交换并捕捉跨模态全局依赖性。采用多头注意力进一步增强特征表示并提高模型描绘医学实体边界的能力。条件随机场(CRF)层用于解码,确保实体预测的全局一致性,从而提高识别准确性和鲁棒性。所提出的方法在CMeEE-V2、IMCS-V2-NER和CHIP-STS数据集上分别取得了65.26%、80.31%和86.73%的F1分数,优于其他模型,并在医学实体识别准确性和多个评估指标上显示出显著提高。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec59/12157088/f1ed6390bb7c/pone.0325660.g001.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验