• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于 ERNIE 模型迁移学习的知识蒸馏在智能对话意图识别中的应用。

Application of Knowledge Distillation Based on Transfer Learning of ERNIE Model in Intelligent Dialogue Intention Recognition.

机构信息

School of Information & Communication Engineering, Beijing Information Science and Technology University, Beijing 100025, China.

出版信息

Sensors (Basel). 2022 Feb 8;22(3):1270. doi: 10.3390/s22031270.

DOI:10.3390/s22031270
PMID:35162015
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8838728/
Abstract

The 'intention' classification of a user question is an important element of a task-engine driven chatbot. The essence of a user question's intention understanding is the text classification. The transfer learning, such as BERT (Bidirectional Encoder Representations from Transformers) and ERNIE (Enhanced Representation through Knowledge Integration), has put the text classification task into a new level, but the BERT and ERNIE model are difficult to support high QPS (queries per second) intelligent dialogue systems due to computational performance issues. In reality, the simple classification model usually shows a high computational performance, but they are limited by low accuracy. In this paper, we use knowledge of the ERNIE model to distill the FastText model; the ERNIE model works as a teacher model to predict the massive online unlabeled data for data enhancement, and then guides the training of the student model of FastText with better computational efficiency. The FastText model is distilled by the ERNIE model in chatbot intention classification. This not only guarantees the superiority of its original computational performance, but also the intention classification accuracy has been significantly improved.

摘要

用户问题的“意图”分类是任务引擎驱动的聊天机器人的重要元素。用户问题意图理解的本质是文本分类。迁移学习,如 BERT(来自转换器的双向编码器表示)和 ERNIE(通过知识集成增强表示),将文本分类任务提升到了一个新的水平,但由于计算性能问题,BERT 和 ERNIE 模型难以支持高 QPS(每秒查询量)智能对话系统。实际上,简单的分类模型通常表现出较高的计算性能,但它们受到准确性较低的限制。在本文中,我们使用 ERNIE 模型的知识来提炼 FastText 模型;ERNIE 模型作为教师模型来预测大规模的在线无标签数据以进行数据增强,然后指导具有更好计算效率的 FastText 学生模型的训练。ERNIE 模型在聊天机器人意图分类中对 FastText 模型进行蒸馏。这不仅保证了其原始计算性能的优越性,而且意图分类的准确性也得到了显著提高。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/a0cc01879303/sensors-22-01270-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/b863c43a70e8/sensors-22-01270-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/2ea4b53140b1/sensors-22-01270-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/869d21d226ee/sensors-22-01270-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/238c47c1c799/sensors-22-01270-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/603581a78043/sensors-22-01270-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/4a6cc70dd0cc/sensors-22-01270-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/accae640889e/sensors-22-01270-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/a0cc01879303/sensors-22-01270-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/b863c43a70e8/sensors-22-01270-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/2ea4b53140b1/sensors-22-01270-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/869d21d226ee/sensors-22-01270-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/238c47c1c799/sensors-22-01270-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/603581a78043/sensors-22-01270-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/4a6cc70dd0cc/sensors-22-01270-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/accae640889e/sensors-22-01270-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0dbb/8838728/a0cc01879303/sensors-22-01270-g008.jpg

相似文献

1
Application of Knowledge Distillation Based on Transfer Learning of ERNIE Model in Intelligent Dialogue Intention Recognition.基于 ERNIE 模型迁移学习的知识蒸馏在智能对话意图识别中的应用。
Sensors (Basel). 2022 Feb 8;22(3):1270. doi: 10.3390/s22031270.
2
Multi-Label Classification in Patient-Doctor Dialogues With the RoBERTa-WWM-ext + CNN (Robustly Optimized Bidirectional Encoder Representations From Transformers Pretraining Approach With Whole Word Masking Extended Combining a Convolutional Neural Network) Model: Named Entity Study.基于RoBERTa-WWM-ext + CNN(带有全词掩码扩展的基于变换器预训练方法的稳健优化双向编码器表示与卷积神经网络相结合)模型的医患对话多标签分类:命名实体研究
JMIR Med Inform. 2022 Apr 21;10(4):e35606. doi: 10.2196/35606.
3
Automated classification of clinical trial eligibility criteria text based on ensemble learning and metric learning.基于集成学习和度量学习的临床试验资格标准文本的自动分类。
BMC Med Inform Decis Mak. 2021 Jul 30;21(Suppl 2):129. doi: 10.1186/s12911-021-01492-z.
4
Deep Learning Approach for Negation and Speculation Detection for Automated Important Finding Flagging and Extraction in Radiology Report: Internal Validation and Technique Comparison Study.用于放射学报告中自动重要发现标记和提取的否定与推测检测的深度学习方法:内部验证与技术比较研究
JMIR Med Inform. 2023 Apr 25;11:e46348. doi: 10.2196/46348.
5
Medical Specialty Recommendations by an Artificial Intelligence Chatbot on a Smartphone: Development and Deployment.智能手机人工智能聊天机器人的医学专业推荐:开发与部署。
J Med Internet Res. 2021 May 6;23(5):e27460. doi: 10.2196/27460.
6
Resolution-based distillation for efficient histology image classification.基于分辨率的蒸馏用于高效的组织学图像分类。
Artif Intell Med. 2021 Sep;119:102136. doi: 10.1016/j.artmed.2021.102136. Epub 2021 Aug 6.
7
Identifying the Question Similarity of Regulatory Documents in the Pharmaceutical Industry by Using the Recognizing Question Entailment System: Evaluation Study.利用识别问题蕴含系统识别制药行业监管文件中的问题相似性:评估研究
JMIR AI. 2023 Sep 26;2:e43483. doi: 10.2196/43483.
8
Sentiment analysis of hotel online reviews using the BERT model and ERNIE model-Data from China.基于 BERT 模型和 ERNIE 模型的酒店在线评论情感分析——来自中国的数据。
PLoS One. 2023 Mar 10;18(3):e0275382. doi: 10.1371/journal.pone.0275382. eCollection 2023.
9
Medical text classification based on the discriminative pre-training model and prompt-tuning.基于判别式预训练模型和提示调整的医学文本分类
Digit Health. 2023 Aug 6;9:20552076231193213. doi: 10.1177/20552076231193213. eCollection 2023 Jan-Dec.
10
Transfer Learning for Sentiment Classification Using Bidirectional Encoder Representations from Transformers (BERT) Model.使用来自Transformer的双向编码器表征(BERT)模型进行情感分类的迁移学习
Sensors (Basel). 2023 May 31;23(11):5232. doi: 10.3390/s23115232.

引用本文的文献

1
A Comparative Study of Natural Language Processing Algorithms Based on Cities Changing Diabetes Vulnerability Data.基于城市糖尿病易感性数据的自然语言处理算法对比研究
Healthcare (Basel). 2022 Jun 15;10(6):1119. doi: 10.3390/healthcare10061119.

本文引用的文献

1
Long short-term memory.长短期记忆
Neural Comput. 1997 Nov 15;9(8):1735-80. doi: 10.1162/neco.1997.9.8.1735.