• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于深度学习的人机多轮语言对话交互

Human-Machine Multi-Turn Language Dialogue Interaction Based on Deep Learning.

作者信息

Ke Xianxin, Hu Ping, Yang Chenghao, Zhang Renbao

机构信息

School of Mechanical and Electrical Engineering and Automation, Shanghai University, Shanghai 200444, China.

出版信息

Micromachines (Basel). 2022 Feb 23;13(3):355. doi: 10.3390/mi13030355.

DOI:10.3390/mi13030355
PMID:35334647
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8955483/
Abstract

During multi-turn dialogue, with the increase in dialogue turns, the difficulty of intention recognition and the generation of the following sentence reply become more and more difficult. This paper mainly optimizes the context information extraction ability of the Seq2Seq Encoder in multi-turn dialogue modeling. We fuse the historical dialogue information and the current input statement information in the encoder to capture the context dialogue information better. Therefore, we propose a BERT-based fusion encoder ProBERT-To-GUR (PBTG) and an enhanced ELMO model 3-ELMO-Attention-GRU (3EAG). The two models mainly enhance the contextual information extraction capability of multi-turn dialogue. To verify the effectiveness of the two proposed models, we demonstrate the effectiveness of our model by combining data based on the LCCC-large multi-turn dialogue dataset and the Naturalconv multi-turn dataset. The experimental comparison results show that, in the multi-turn dialogue experiments of the open domain and fixed topic, the two Seq2Seq coding models proposed are significantly improved compared with the current state-of-the-art models. For specified topic multi-turn dialogue, the 3EAG model has the average BLEU value reaches the optimal 32.4, which achieves the best language generation effect, and the BLEU value in the actual dialogue verification experiment also surpasses 31.8. for open-domain multi-turn dialogue. The average BLEU value of the PBTG model reaches 31.8, the optimal 31.8 achieves the best language generation effect, and the BLEU value in the actual dialogue verification experiment surpasses 31.2. So, the 3EAG model is more suitable for fixed-topic multi-turn dialogues for the two tasks. The PBTG model is more muscular in open-domain multi-turn dialogue tasks; therefore, our model is significant for promoting multi-turn dialogue research.

摘要

在多轮对话中,随着对话轮数的增加,意图识别的难度以及生成后续句子回复变得越来越困难。本文主要优化了多轮对话建模中Seq2Seq编码器的上下文信息提取能力。我们在编码器中融合历史对话信息和当前输入语句信息,以更好地捕捉上下文对话信息。因此,我们提出了基于BERT的融合编码器ProBERT-To-GUR(PBTG)和增强的ELMO模型3-ELMO-Attention-GRU(3EAG)。这两个模型主要增强了多轮对话的上下文信息提取能力。为了验证所提出的两个模型的有效性,我们通过结合基于LCCC-large多轮对话数据集和Naturalconv多轮数据集的数据来证明我们模型的有效性。实验比较结果表明,在开放域和固定主题的多轮对话实验中,所提出的两个Seq2Seq编码模型与当前最先进的模型相比有显著改进。对于指定主题的多轮对话,3EAG模型的平均BLEU值达到最优的32.4,实现了最佳的语言生成效果,并且在实际对话验证实验中的BLEU值也超过了31.8。对于开放域多轮对话,PBTG模型的平均BLEU值达到31.8,最优的31.8实现了最佳的语言生成效果,并且在实际对话验证实验中的BLEU值超过了31.2。所以,对于这两项任务,3EAG模型更适合固定主题的多轮对话。PBTG模型在开放域多轮对话任务中更有优势;因此,我们的模型对推动多轮对话研究具有重要意义。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7094/8955483/6e7755254d6c/micromachines-13-00355-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7094/8955483/85589f2512a9/micromachines-13-00355-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7094/8955483/743d49b9fe9f/micromachines-13-00355-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7094/8955483/8dcb187010aa/micromachines-13-00355-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7094/8955483/5a97137313d9/micromachines-13-00355-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7094/8955483/4fb11062490f/micromachines-13-00355-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7094/8955483/6e7755254d6c/micromachines-13-00355-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7094/8955483/85589f2512a9/micromachines-13-00355-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7094/8955483/743d49b9fe9f/micromachines-13-00355-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7094/8955483/8dcb187010aa/micromachines-13-00355-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7094/8955483/5a97137313d9/micromachines-13-00355-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7094/8955483/4fb11062490f/micromachines-13-00355-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7094/8955483/6e7755254d6c/micromachines-13-00355-g006.jpg

相似文献

1
Human-Machine Multi-Turn Language Dialogue Interaction Based on Deep Learning.基于深度学习的人机多轮语言对话交互
Micromachines (Basel). 2022 Feb 23;13(3):355. doi: 10.3390/mi13030355.
2
EAGS: An extracting auxiliary knowledge graph model in multi-turn dialogue generation.EAGS:多轮对话生成中的一种提取辅助知识图谱模型。
World Wide Web. 2022 Sep 30:1-22. doi: 10.1007/s11280-022-01100-8.
3
Extracting comprehensive clinical information for breast cancer using deep learning methods.利用深度学习方法提取乳腺癌全面临床信息。
Int J Med Inform. 2019 Dec;132:103985. doi: 10.1016/j.ijmedinf.2019.103985. Epub 2019 Oct 2.
4
HQA-Data: A historical question answer generation dataset from previous multi perspective conversation.HQA-数据:一个来自以往多视角对话的历史问答生成数据集。
Data Brief. 2023 May 18;48:109245. doi: 10.1016/j.dib.2023.109245. eCollection 2023 Jun.
5
Multiturn dialogue generation by modeling sentence-level and discourse-level contexts.通过对句子级和语篇级上下文进行建模实现多轮对话生成。
Sci Rep. 2022 Nov 27;12(1):20349. doi: 10.1038/s41598-022-24787-1.
6
Enhancing Clinical Accuracy of Medical Chatbots with Large Language Models.利用大语言模型提高医疗聊天机器人的临床准确性。
IEEE J Biomed Health Inform. 2024 Sep 27;PP. doi: 10.1109/JBHI.2024.3470323.
7
Topic Break Detection in Interview Dialogues Using Sentence Embedding of Utterance and Speech Intention Based on Multitask Neural Networks.基于多任务神经网络的话语句子嵌入和语音意图的访谈对话主题中断检测。
Sensors (Basel). 2022 Jan 17;22(2):694. doi: 10.3390/s22020694.
8
End-to-End Dialogue Generation Using a Single Encoder and a Decoder Cascade With a Multidimension Attention Mechanism.使用带有多维注意力机制的单个编码器和解码器级联进行端到端对话生成
IEEE Trans Neural Netw Learn Syst. 2023 Nov;34(11):8482-8492. doi: 10.1109/TNNLS.2022.3151347. Epub 2023 Oct 27.
9
Multi-Level Representation Learning for Chinese Medical Entity Recognition: Model Development and Validation.用于中文医学实体识别的多层次表示学习:模型开发与验证
JMIR Med Inform. 2020 May 4;8(5):e17637. doi: 10.2196/17637.
10
Comparing deep learning architectures for sentiment analysis on drug reviews.比较药物评论情感分析的深度学习架构。
J Biomed Inform. 2020 Oct;110:103539. doi: 10.1016/j.jbi.2020.103539. Epub 2020 Aug 17.