基于对抗性知识蒸馏的生物医学事实问答。
Adversarial Knowledge Distillation Based Biomedical Factoid Question Answering.
出版信息
IEEE/ACM Trans Comput Biol Bioinform. 2023 Jan-Feb;20(1):106-118. doi: 10.1109/TCBB.2022.3161032. Epub 2023 Feb 3.
Biomedical factoid question answering is an essential application for biomedical information sharing. Recently, neural network based approaches have shown remarkable performance for this task. However, due to the scarcity of annotated data which requires intensive knowledge of expertise, training a robust model on limited-scale biomedical datasets remains a challenge. Previous works solve this problem by introducing useful knowledge. It is found that the interaction between question and answer (QA-interaction) is also a kind of knowledge which could help extract answer accurately. This research develops a knowledge distillation framework for biomedical factoid question answering, in which a teacher model as the knowledge source of QA-interaction is designed to enhance the student model. In addition, to further alleviate the problem of limited-scale dataset, a novel adversarial knowledge distillation technique is proposed to robustly distill the knowledge from teacher model to student model by constructing perturbed examples as additional training data. By forcing the student model to mimic the predicted distributions of teacher model on both original examples and perturbed examples, the knowledge of QA-interaction can be learned by student model. We evaluate the proposed framework on the widely used BioASQ datasets, and experimental results have shown the proposed method's promising potential.
生物医学事实问答是生物医学信息共享的重要应用。最近,基于神经网络的方法在这项任务中表现出了显著的性能。然而,由于需要专业知识的注释数据稀缺,在有限规模的生物医学数据集上训练稳健的模型仍然是一个挑战。以前的工作通过引入有用的知识来解决这个问题。研究发现,问题和答案之间的交互(QA 交互)也是一种可以帮助准确提取答案的知识。本研究开发了一种用于生物医学事实问答的知识蒸馏框架,其中设计了一个教师模型作为 QA 交互的知识源,以增强学生模型。此外,为了进一步缓解有限规模数据集的问题,提出了一种新颖的对抗性知识蒸馏技术,通过构建扰动示例作为额外的训练数据,从教师模型中稳健地提取知识到学生模型。通过迫使学生模型模仿教师模型在原始示例和扰动示例上的预测分布,学生模型可以学习 QA 交互的知识。我们在广泛使用的 BioASQ 数据集上评估了所提出的框架,实验结果表明了该方法的有前途的潜力。