• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

使用多通道卷积神经网络-双向长短期记忆网络进行联合意图检测和带有句法与语义特征的槽填充

Joint intent detection and slot filling with syntactic and semantic features using multichannel CNN-BiLSTM.

作者信息

Muhammad Yusuf Idris, Salim Naomie, Zainal Anazida

机构信息

Faculty of Computing, Universiti Teknologi Malaysia, Skudai, Johor, Malaysia.

出版信息

PeerJ Comput Sci. 2024 Oct 8;10:e2346. doi: 10.7717/peerj-cs.2346. eCollection 2024.

DOI:10.7717/peerj-cs.2346
PMID:39650516
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11623216/
Abstract

Understanding spoken language is crucial for conversational agents, with intent detection and slot filling being the primary tasks in natural language understanding (NLU). Enhancing the NLU tasks can lead to an accurate and efficient virtual assistant thereby reducing the need for human intervention and expanding their applicability in other domains. Traditionally, these tasks have been addressed individually, but recent studies have highlighted their interconnection, suggesting better results when solved together. Recent advances in natural language processing have shown that pretrained word embeddings can enhance text representation and improve the generalization capabilities of models. However, the challenge of poor generalization in joint learning models for intent detection and slot filling remains due to limited annotated datasets. Additionally, traditional models face difficulties in capturing both the semantic and syntactic nuances of language, which are vital for accurate intent detection and slot filling. This study proposes a hybridized text representation method using a multichannel convolutional neural network with three embedding channels: non-contextual embeddings for semantic information, part-of-speech (POS) tag embeddings for syntactic features, and contextual embeddings for deeper contextual understanding. Specifically, we utilized word2vec for non-contextual embeddings, one-hot vectors for POS tags, and bidirectional encoder representations from transformers (BERT) for contextual embeddings. These embeddings are processed through a convolutional layer and a shared bidirectional long short-term memory (BiLSTM) network, followed by two softmax functions for intent detection and slot filling. Experiments on the air travel information system (ATIS) and SNIPS datasets demonstrated that our model significantly outperformed the baseline models, achieving an intent accuracy of 97.90% and slot filling F1-score of 98.86% on the ATIS dataset, and an intent accuracy of 98.88% and slot filling F1-score of 97.07% on the SNIPS dataset. These results highlight the effectiveness of our proposed approach in advancing dialogue systems, and paving the way for more accurate and efficient natural language understanding in real-world applications.

摘要

对于对话智能体而言,理解口语至关重要,意图检测和槽位填充是自然语言理解(NLU)中的主要任务。增强NLU任务能够带来准确且高效的虚拟助手,从而减少对人工干预的需求,并扩大其在其他领域的适用性。传统上,这些任务是分别处理的,但最近的研究强调了它们之间的相互联系,表明一起解决时会取得更好的结果。自然语言处理的最新进展表明,预训练词嵌入可以增强文本表示并提高模型的泛化能力。然而,由于标注数据集有限,意图检测和槽位填充联合学习模型中泛化能力差的挑战依然存在。此外,传统模型在捕捉语言的语义和句法细微差别方面面临困难,而这些细微差别对于准确的意图检测和槽位填充至关重要。本研究提出了一种混合文本表示方法,该方法使用具有三个嵌入通道的多通道卷积神经网络:用于语义信息的非上下文嵌入、用于句法特征的词性(POS)标签嵌入以及用于更深入上下文理解的上下文嵌入。具体而言,我们使用word2vec进行非上下文嵌入,使用独热向量表示POS标签,并使用来自变换器的双向编码器表示(BERT)进行上下文嵌入。这些嵌入通过一个卷积层和一个共享的双向长短期记忆(BiLSTM)网络进行处理,随后通过两个softmax函数进行意图检测和槽位填充。在航空旅行信息系统(ATIS)和SNIPS数据集上的实验表明,我们的模型显著优于基线模型,在ATIS数据集上实现了97.90%的意图准确率和98.86%的槽位填充F1分数,在SNIPS数据集上实现了98.88%的意图准确率和97.07%的槽位填充F1分数。这些结果凸显了我们提出方法在推进对话系统方面的有效性,并为在实际应用中实现更准确高效的自然语言理解铺平了道路。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/99c2/11623216/54485bb91642/peerj-cs-10-2346-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/99c2/11623216/d58d22b5727b/peerj-cs-10-2346-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/99c2/11623216/a82aa1e5538c/peerj-cs-10-2346-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/99c2/11623216/b5fc75e061c6/peerj-cs-10-2346-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/99c2/11623216/939586e6afee/peerj-cs-10-2346-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/99c2/11623216/54485bb91642/peerj-cs-10-2346-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/99c2/11623216/d58d22b5727b/peerj-cs-10-2346-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/99c2/11623216/a82aa1e5538c/peerj-cs-10-2346-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/99c2/11623216/b5fc75e061c6/peerj-cs-10-2346-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/99c2/11623216/939586e6afee/peerj-cs-10-2346-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/99c2/11623216/54485bb91642/peerj-cs-10-2346-g005.jpg

相似文献

1
Joint intent detection and slot filling with syntactic and semantic features using multichannel CNN-BiLSTM.使用多通道卷积神经网络-双向长短期记忆网络进行联合意图检测和带有句法与语义特征的槽填充
PeerJ Comput Sci. 2024 Oct 8;10:e2346. doi: 10.7717/peerj-cs.2346. eCollection 2024.
2
Pre-Trained Joint Model for Intent Classification and Slot Filling with Semantic Feature Fusion.基于语义特征融合的意图分类和槽位填充的预训练联合模型。
Sensors (Basel). 2023 Mar 6;23(5):2848. doi: 10.3390/s23052848.
3
A Two-Stage Selective Fusion Framework for Joint Intent Detection and Slot Filling.一种用于联合意图检测和槽填充的两阶段选择性融合框架。
IEEE Trans Neural Netw Learn Syst. 2024 Mar;35(3):3874-3885. doi: 10.1109/TNNLS.2022.3202562. Epub 2024 Feb 29.
4
Natural language understanding of map navigation queries in Roman Urdu by joint entity and intent determination.通过联合实体和意图确定实现对乌尔都语罗马文地图导航查询的自然语言理解。
PeerJ Comput Sci. 2021 Jul 21;7:e615. doi: 10.7717/peerj-cs.615. eCollection 2021.
5
A Study on the Impacts of Slot Types and Training Data on Joint Natural Language Understanding in a Spanish Medication Management Assistant Scenario.西班牙药物管理助手场景下插槽类型和训练数据对联合自然语言理解的影响研究
Sensors (Basel). 2022 Mar 18;22(6):2364. doi: 10.3390/s22062364.
6
BioBBC: a multi-feature model that enhances the detection of biomedical entities.生物 BBC:一种增强生物医学实体检测的多特征模型。
Sci Rep. 2024 Apr 2;14(1):7697. doi: 10.1038/s41598-024-58334-x.
7
Multi-Label Classification in Patient-Doctor Dialogues With the RoBERTa-WWM-ext + CNN (Robustly Optimized Bidirectional Encoder Representations From Transformers Pretraining Approach With Whole Word Masking Extended Combining a Convolutional Neural Network) Model: Named Entity Study.基于RoBERTa-WWM-ext + CNN(带有全词掩码扩展的基于变换器预训练方法的稳健优化双向编码器表示与卷积神经网络相结合)模型的医患对话多标签分类:命名实体研究
JMIR Med Inform. 2022 Apr 21;10(4):e35606. doi: 10.2196/35606.
8
An efficient method for disaster tweets classification using gradient-based optimized convolutional neural networks with BERT embeddings.一种使用基于梯度优化的卷积神经网络与BERT嵌入的高效灾难推文分类方法。
MethodsX. 2024 Jul 3;13:102843. doi: 10.1016/j.mex.2024.102843. eCollection 2024 Dec.
9
The Impact of Pretrained Language Models on Negation and Speculation Detection in Cross-Lingual Medical Text: Comparative Study.预训练语言模型对跨语言医学文本中否定和推测检测的影响:比较研究
JMIR Med Inform. 2020 Dec 3;8(12):e18953. doi: 10.2196/18953.
10
Enhancing clinical concept extraction with contextual embeddings.利用上下文嵌入增强临床概念提取。
J Am Med Inform Assoc. 2019 Nov 1;26(11):1297-1304. doi: 10.1093/jamia/ocz096.

引用本文的文献

1
Detecting sarcasm in user-generated content integrating transformers and gated graph neural networks.利用Transformer和门控图神经网络检测用户生成内容中的讽刺意味。
PeerJ Comput Sci. 2025 Apr 21;11:e2817. doi: 10.7717/peerj-cs.2817. eCollection 2025.

本文引用的文献

1
FLMatchQA: a recursive neural network-based question answering with customized federated learning model.FLMatchQA:一种基于递归神经网络的问答系统,采用定制的联邦学习模型。
PeerJ Comput Sci. 2024 Jun 28;10:e2092. doi: 10.7717/peerj-cs.2092. eCollection 2024.
2
Daily natural gas load prediction method based on APSO optimization and Attention-BiLSTM.基于APSO优化与注意力双向长短期记忆网络的每日天然气负荷预测方法
PeerJ Comput Sci. 2024 Feb 29;10:e1890. doi: 10.7717/peerj-cs.1890. eCollection 2024.
3
Exploring Convolutional Neural Network Architectures for EEG Feature Extraction.
探索卷积神经网络架构在 EEG 特征提取中的应用。
Sensors (Basel). 2024 Jan 29;24(3):877. doi: 10.3390/s24030877.
4
A improved pooling method for convolutional neural networks.一种用于卷积神经网络的改进池化方法。
Sci Rep. 2024 Jan 18;14(1):1589. doi: 10.1038/s41598-024-51258-6.