Sun Yuanyuan, Gao Dongping, Shen Xifeng, Li Meiting, Nan Jiale, Zhang Weining
Institute of Medical Information, Chinese Academy of Medical Sciences, Peking Union Medical College, Beijing, China.
Department of Internal Medicine, Chinese Academy of Medical Sciences, Peking Union Medical College Hospital, Beijing, China.
JMIR Med Inform. 2022 Apr 21;10(4):e35606. doi: 10.2196/35606.
With the prevalence of online consultation, many patient-doctor dialogues have accumulated, which, in an authentic language environment, are of significant value to the research and development of intelligent question answering and automated triage in recent natural language processing studies.
The purpose of this study was to design a front-end task module for the network inquiry of intelligent medical services. Through the study of automatic labeling of real doctor-patient dialogue text on the internet, a method of identifying the negative and positive entities of dialogues with higher accuracy has been explored.
The data set used for this study was from the Spring Rain Doctor internet online consultation, which was downloaded from the official data set of Alibaba Tianchi Lab. We proposed a composite abutting joint model, which was able to automatically classify the types of clinical finding entities into the following 4 attributes: positive, negative, other, and empty. We adapted a downstream architecture in Chinese Robustly Optimized Bidirectional Encoder Representations from Transformers Pretraining Approach (RoBERTa) with whole word masking (WWM) extended (RoBERTa-WWM-ext) combining a text convolutional neural network (CNN). We used RoBERTa-WWM-ext to express sentence semantics as a text vector and then extracted the local features of the sentence through the CNN, which was our new fusion model. To verify its knowledge learning ability, we chose Enhanced Representation through Knowledge Integration (ERNIE), original Bidirectional Encoder Representations from Transformers (BERT), and Chinese BERT with WWM to perform the same task, and then compared the results. Precision, recall, and macro-F1 were used to evaluate the performance of the methods.
We found that the ERNIE model, which was trained with a large Chinese corpus, had a total score (macro-F1) of 65.78290014, while BERT and BERT-WWM had scores of 53.18247117 and 69.2795315, respectively. Our composite abutting joint model (RoBERTa-WWM-ext + CNN) had a macro-F1 value of 70.55936311, showing that our model outperformed the other models in the task.
The accuracy of the original model can be greatly improved by giving priority to WWM and replacing the word-based mask with unit to classify and label medical entities. Better results can be obtained by effectively optimizing the downstream tasks of the model and the integration of multiple models later on. The study findings contribute to the translation of online consultation information into machine-readable information.
随着在线问诊的普及,积累了大量医患对话,在真实语言环境下,这些对话对近期自然语言处理研究中的智能问答和自动分诊的研发具有重要价值。
本研究旨在设计一个智能医疗服务网络问诊的前端任务模块。通过对互联网上真实医患对话文本的自动标注研究,探索一种能更准确识别对话中正负实体的方法。
本研究使用的数据集来自春雨医生互联网在线问诊,从阿里巴巴天池实验室的官方数据集中下载。我们提出了一种复合邻接模型,该模型能够将临床发现实体的类型自动分类为以下4种属性:阳性、阴性、其他和空。我们采用了一种下游架构,即带有全词掩码扩展的中文稳健优化双向编码器表征预训练方法(RoBERTa)(RoBERTa-WWM-ext)与文本卷积神经网络(CNN)相结合。我们使用RoBERTa-WWM-ext将句子语义表示为文本向量,然后通过CNN提取句子的局部特征,这就是我们的新融合模型。为验证其知识学习能力,我们选择通过知识整合增强表征(ERNIE)、原始的双向编码器表征来自Transformer(BERT)以及带有全词掩码的中文BERT来执行相同任务,然后比较结果。精确率、召回率和宏F1用于评估这些方法的性能。
我们发现,使用大量中文语料库训练的ERNIE模型的总分(宏F1)为65.78290014,而BERT和BERT-WWM的分数分别为53.18247117和69.2795315。我们的复合邻接模型(RoBERTa-WWM-ext + CNN)的宏F1值为70.55936311,表明我们的模型在该任务中优于其他模型。
通过优先使用全词掩码并将基于词的掩码替换为单元来对医学实体进行分类和标注,可以大大提高原始模型的准确性。通过有效优化模型的下游任务以及随后整合多个模型,可以获得更好的结果。研究结果有助于将在线问诊信息转化为机器可读信息。