Hou Yutai, Wang Xinghao, Chen Cheng, Li Bohan, Che Wanxiang, Chen Zhigang
Research Center for Social Computing and Information Retrieval, Harbin Institute of Technology, Harbin, China.
State Key Laboratory of Cognitive Intelligence, Hefei, China.
Int J Mach Learn Cybern. 2022;13(11):3409-3423. doi: 10.1007/s13042-022-01604-9. Epub 2022 Jul 18.
Few-shot learning (FSL) is one of the key future steps in machine learning and raises a lot of attention. In this paper, we focus on the FSL problem of dialogue understanding, which contains two closely related tasks: intent detection and slot filling. Dialogue understanding has been proven to benefit a lot from jointly learning the two sub-tasks. However, such joint learning becomes challenging in the few-shot scenarios: on the one hand, the sparsity of samples greatly magnifies the difficulty of modeling the connection between the two tasks; on the other hand, how to jointly learn multiple tasks in the few-shot setting is still less investigated. In response to this, we introduce FewJoint, the first FSL benchmark for joint dialogue understanding. FewJoint provides a new corpus with 59 different dialogue domains from real industrial API and a code platform to ease FSL experiment set-up, which are expected to advance the research of this field. Further, we find that insufficient performance of the few-shot setting often leads to noisy sharing between two sub-task and disturbs joint learning. To tackle this, we guide slot with explicit intent information and propose a novel trust gating mechanism that blocks low-confidence intent information to ensure high quality sharing. Besides, we introduce a Reptile-based meta-learning strategy to achieve better generalization in unseen few-shot domains. In the experiments, the proposed method brings significant improvements on two datasets and achieve new state-of-the-art performance.
少样本学习(FSL)是机器学习未来的关键步骤之一,引起了广泛关注。在本文中,我们关注对话理解的少样本学习问题,它包含两个密切相关的任务:意图检测和槽填充。事实证明,对话理解通过联合学习这两个子任务受益匪浅。然而,在少样本场景中,这种联合学习变得具有挑战性:一方面,样本的稀疏性极大地增加了对两个任务之间联系进行建模的难度;另一方面,在少样本设置下如何联合学习多个任务仍较少被研究。针对此,我们引入了FewJoint,这是首个用于联合对话理解的少样本学习基准。FewJoint提供了一个新的语料库,包含来自真实工业应用程序编程接口的59个不同对话领域以及一个代码平台,以简化少样本学习实验设置,有望推动该领域的研究。此外,我们发现少样本设置下性能不足通常会导致两个子任务之间的噪声共享,并干扰联合学习。为解决此问题,我们用明确的意图信息引导槽填充,并提出一种新颖的信任门控机制,该机制会阻止低置信度的意图信息,以确保高质量的共享。此外,我们引入一种基于Reptile的元学习策略,以在未见过的少样本领域中实现更好的泛化。在实验中,所提出的方法在两个数据集上带来了显著改进,并取得了新的最优性能。