• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

ICMC:一种用于教学计划评分的可解释跨域多模态分类模型。

ICMC: An Interpretable Cross-domain Multi-modal Classification model for grading teaching plan.

作者信息

Jin Jin, Wang Fan, Tian Shengzheng

机构信息

School of Information and Intelligent Engineering, Zhejiang Wanli University, Ningbo, Zhejiang, China.

ZHONGTIETONG Rail Transit Operation Co. Ltd., Wenzhou, Zhejiang, China.

出版信息

PLoS One. 2025 Sep 3;20(9):e0330684. doi: 10.1371/journal.pone.0330684. eCollection 2025.

DOI:10.1371/journal.pone.0330684
PMID:40901896
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12407483/
Abstract

Multi-modal classification aims to extract pertinent information from various modalities to assign labels to instances. The advent of deep neural networks has significantly advanced this task. However, the majority of current deep neural networks lack interpretability, leading to skepticism. This issue is particularly pronounced in sensitive domains such as educational assessment. In order to address the trust deficit in deep neural networks for multi-modal classification tasks, we propose an Interpretable Multi-modal Classification framework (ICMC), which enhances confidence in the processes and outcomes of deep neural networks while maintaining interpretability and improving performance. Specifically, our approach incorporates a confidence-driven attention mechanism at the intermediate layer of the deep neural network, assessing attention scores and discerning anomalous information from both local and global perspectives. Furthermore, a confidence probability mechanism is implemented at the output layer, leveraging both local and global perspectives to bolster result confidence. Additionally, we meticulously curate multi-modal datasets for automatic lesson plan scoring research, making them openly available to the community. Quantitative experiments on educational and medical datasets confirm that ICMC outperforms state-of-the-art models (HMCAN, MCAN, HGLNet) by 2.5-6.0% in accuracy and 3.1-7.2% in F1-score, while reducing computational latency by 18%. Cross-domain validation demonstrates 15.7% higher generalizability than transformer-based approaches (CLIP), establishing its interpretability through attention visualization and confidence scoring.

摘要

多模态分类旨在从各种模态中提取相关信息,以便为实例分配标签。深度神经网络的出现显著推动了这项任务的发展。然而,当前大多数深度神经网络缺乏可解释性,引发了质疑。这个问题在教育评估等敏感领域尤为突出。为了解决多模态分类任务中深度神经网络的信任赤字问题,我们提出了一个可解释的多模态分类框架(ICMC),该框架在保持可解释性并提高性能的同时,增强了对深度神经网络过程和结果的信心。具体而言,我们的方法在深度神经网络的中间层引入了一种置信度驱动的注意力机制,从局部和全局角度评估注意力分数并识别异常信息。此外,在输出层实现了一种置信概率机制,利用局部和全局视角来增强结果的置信度。此外,我们精心策划了用于自动教案评分研究的多模态数据集,并向社区公开提供这些数据集。在教育和医疗数据集上的定量实验证实,ICMC在准确率上比现有最先进的模型(HMCAN、MCAN、HGLNet)高出2.5 - 6.0%,在F1分数上高出3.1 - 7.2%,同时将计算延迟降低了18%。跨域验证表明,其泛化能力比基于Transformer的方法(CLIP)高15.7%,通过注意力可视化和置信度评分确立了其可解释性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/c86c392f4975/pone.0330684.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/fd78f0d30527/pone.0330684.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/6fb8cfa03ef0/pone.0330684.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/7c56b30ae02a/pone.0330684.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/8b0addda5e97/pone.0330684.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/da09c879f949/pone.0330684.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/6d6470bb8568/pone.0330684.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/c86c392f4975/pone.0330684.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/fd78f0d30527/pone.0330684.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/6fb8cfa03ef0/pone.0330684.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/7c56b30ae02a/pone.0330684.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/8b0addda5e97/pone.0330684.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/da09c879f949/pone.0330684.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/6d6470bb8568/pone.0330684.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dd60/12407483/c86c392f4975/pone.0330684.g007.jpg

相似文献

1
ICMC: An Interpretable Cross-domain Multi-modal Classification model for grading teaching plan.ICMC:一种用于教学计划评分的可解释跨域多模态分类模型。
PLoS One. 2025 Sep 3;20(9):e0330684. doi: 10.1371/journal.pone.0330684. eCollection 2025.
2
Prescription of Controlled Substances: Benefits and Risks管制药品的处方:益处与风险
3
CXR-MultiTaskNet a unified deep learning framework for joint disease localization and classification in chest radiographs.CXR-MultiTaskNet:一种用于胸部X光片中疾病联合定位与分类的统一深度学习框架。
Sci Rep. 2025 Aug 31;15(1):32022. doi: 10.1038/s41598-025-16669-z.
4
Short-Term Memory Impairment短期记忆障碍
5
SG-Fusion: A swin-transformer and graph convolution-based multi-modal deep neural network for glioma prognosis.SG-Fusion:一种基于 Swin-Transformer 和图卷积的多模态深度神经网络,用于脑胶质瘤预后。
Artif Intell Med. 2024 Nov;157:102972. doi: 10.1016/j.artmed.2024.102972. Epub 2024 Aug 31.
6
A deep learning approach to direct immunofluorescence pattern recognition in autoimmune bullous diseases.深度学习方法在自身免疫性大疱性疾病中的直接免疫荧光模式识别。
Br J Dermatol. 2024 Jul 16;191(2):261-266. doi: 10.1093/bjd/ljae142.
7
A CrossMod-Transformer deep learning framework for multi-modal pain detection through EDA and ECG fusion.一种用于通过皮肤电活动(EDA)和心电图(ECG)融合进行多模态疼痛检测的CrossMod-Transformer深度学习框架。
Sci Rep. 2025 Aug 12;15(1):29467. doi: 10.1038/s41598-025-14238-y.
8
Deep learning-based image classification for AI-assisted integration of pathology and radiology in medical imaging.基于深度学习的图像分类,用于医学成像中病理学与放射学的人工智能辅助整合。
Front Med (Lausanne). 2025 Jun 2;12:1574514. doi: 10.3389/fmed.2025.1574514. eCollection 2025.
9
iACP-DPNet: a dual-pooling causal dilated convolutional network for interpretable anticancer peptide identification.iACP-DPNet:一种用于可解释抗癌肽识别的双池因果扩张卷积网络。
Funct Integr Genomics. 2025 Jul 4;25(1):147. doi: 10.1007/s10142-025-01641-x.
10
Personalized health monitoring using explainable AI: bridging trust in predictive healthcare.使用可解释人工智能的个性化健康监测:弥合对预测性医疗保健的信任差距。
Sci Rep. 2025 Aug 29;15(1):31892. doi: 10.1038/s41598-025-15867-z.

本文引用的文献

1
HECLIP: histology-enhanced contrastive learning for imputation of transcriptomics profiles.HECLIP:用于转录组学图谱插补的组织学增强对比学习
Bioinformatics. 2025 Jul 1;41(7). doi: 10.1093/bioinformatics/btaf363.
2
Gene expression prediction from histology images via hypergraph neural networks.基于超图神经网络的组织学图像基因表达预测。
Brief Bioinform. 2024 Sep 23;25(6). doi: 10.1093/bib/bbae500.
3
AntiFormer: graph enhanced large language model for binding affinity prediction.AntiFormer:用于结合亲和力预测的图增强大型语言模型。
Brief Bioinform. 2024 Jul 25;25(5). doi: 10.1093/bib/bbae403.
4
Evaluate teaching quality of physical education using a hybrid multi-criteria decision-making framework.运用混合多准则决策框架评估体育教育教学质量。
PLoS One. 2023 Feb 16;18(2):e0280845. doi: 10.1371/journal.pone.0280845. eCollection 2023.
5
Trusted Multi-View Classification With Dynamic Evidential Fusion.基于动态证据融合的可信多视图分类
IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):2551-2566. doi: 10.1109/TPAMI.2022.3171983. Epub 2023 Jan 6.
6
Presentation and evaluation of the teaching concept "ENHANCE" for basic sciences in medical education.呈现和评估医学教育基础科学的教学理念“ENHANCE”。
PLoS One. 2020 Sep 29;15(9):e0239928. doi: 10.1371/journal.pone.0239928. eCollection 2020.
7
Quantitative and qualitative evaluation of a learning model based on workstation activities.基于工作站活动的学习模型的定量和定性评估。
PLoS One. 2020 Aug 5;15(8):e0236940. doi: 10.1371/journal.pone.0236940. eCollection 2020.
8
Mitigating gender bias in student evaluations of teaching.减少教学评估中学生评价中的性别偏见。
PLoS One. 2019 May 15;14(5):e0216241. doi: 10.1371/journal.pone.0216241. eCollection 2019.
9
Advances in cognitive theory and therapy: the generic cognitive model.认知理论与治疗的进展:通用认知模型。
Annu Rev Clin Psychol. 2014;10:1-24. doi: 10.1146/annurev-clinpsy-032813-153734. Epub 2014 Jan 2.