• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于数据分类的具有熵正则化的专家混合模型

Mixture of Experts with Entropic Regularization for Data Classification.

作者信息

Peralta Billy, Saavedra Ariel, Caro Luis, Soto Alvaro

机构信息

Department of Engineering Science, Andres Bello University, Santiago 7500971, Chile.

Department of Engineering Informatics, Catholic University of Temuco, Temuco 4781312, Chile.

出版信息

Entropy (Basel). 2019 Feb 18;21(2):190. doi: 10.3390/e21020190.

DOI:10.3390/e21020190
PMID:33266905
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7514672/
Abstract

Today, there is growing interest in the automatic classification of a variety of tasks, such as weather forecasting, product recommendations, intrusion detection, and people recognition. "Mixture-of-experts" is a well-known classification technique; it is a probabilistic model consisting of local expert classifiers weighted by a gate network that is typically based on softmax functions, combined with learnable complex patterns in data. In this scheme, one data point is influenced by only one expert; as a result, the training process can be misguided in real datasets for which complex data need to be explained by multiple experts. In this work, we propose a variant of the regular mixture-of-experts model. In the proposed model, the cost classification is penalized by the Shannon entropy of the gating network in order to avoid a "winner-takes-all" output for the gating network. Experiments show the advantage of our approach using several real datasets, with improvements in mean accuracy of 3-6% in some datasets. In future work, we plan to embed feature selection into this model.

摘要

如今,人们对各种任务的自动分类越来越感兴趣,比如天气预报、产品推荐、入侵检测和人员识别。“专家混合”是一种著名的分类技术;它是一种概率模型,由基于通常基于softmax函数的门控网络加权的局部专家分类器组成,并结合数据中可学习的复杂模式。在这种方案中,一个数据点仅受一个专家影响;因此,在需要多个专家来解释复杂数据的真实数据集中,训练过程可能会被误导。在这项工作中,我们提出了常规专家混合模型的一个变体。在所提出的模型中,为了避免门控网络出现“赢家通吃”的输出,成本分类由门控网络的香农熵进行惩罚。实验表明,我们的方法在使用几个真实数据集时具有优势,在某些数据集中平均准确率提高了3% - 6%。在未来的工作中,我们计划将特征选择嵌入到这个模型中。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/be89/7514672/5e9396f00384/entropy-21-00190-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/be89/7514672/934a68221ec9/entropy-21-00190-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/be89/7514672/e5a22c1c9593/entropy-21-00190-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/be89/7514672/acb84321f17a/entropy-21-00190-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/be89/7514672/5e9396f00384/entropy-21-00190-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/be89/7514672/934a68221ec9/entropy-21-00190-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/be89/7514672/e5a22c1c9593/entropy-21-00190-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/be89/7514672/acb84321f17a/entropy-21-00190-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/be89/7514672/5e9396f00384/entropy-21-00190-g004.jpg

相似文献

1
Mixture of Experts with Entropic Regularization for Data Classification.用于数据分类的具有熵正则化的专家混合模型
Entropy (Basel). 2019 Feb 18;21(2):190. doi: 10.3390/e21020190.
2
A globally-variant locally-constant model for fusion of labels from multiple diverse experts without using reference labels.一种全球变化局部恒定的模型,用于融合来自多个不同专家的标签,而无需使用参考标签。
IEEE Trans Pattern Anal Mach Intell. 2013 Apr;35(4):769-83. doi: 10.1109/TPAMI.2012.139.
3
Shaping Deep Feature Space Towards Gaussian Mixture for Visual Classification.为视觉分类将深度特征空间塑造为高斯混合模型
IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):2430-2444. doi: 10.1109/TPAMI.2022.3166879. Epub 2023 Jan 6.
4
Entropic One-Class Classifiers.信息嫡一类分类器。
IEEE Trans Neural Netw Learn Syst. 2015 Dec;26(12):3187-200. doi: 10.1109/TNNLS.2015.2418332. Epub 2015 Apr 15.
5
Prediction of paroxysmal Atrial Fibrillation: A machine learning based approach using combined feature vector and mixture of expert classification on HRV signal.基于 HRV 信号的组合特征向量和专家分类混合的机器学习方法预测阵发性心房颤动。
Comput Methods Programs Biomed. 2018 Oct;165:53-67. doi: 10.1016/j.cmpb.2018.07.014. Epub 2018 Aug 10.
6
Using ELM-based weighted probabilistic model in the classification of synchronous EEG BCI.基于极限学习机的加权概率模型在同步脑电图脑机接口分类中的应用
Med Biol Eng Comput. 2017 Jan;55(1):33-43. doi: 10.1007/s11517-016-1493-x. Epub 2016 Apr 21.
7
Mixture of experts for classification of gender, ethnic origin, and pose of human faces.
IEEE Trans Neural Netw. 2000;11(4):948-60. doi: 10.1109/72.857774.
8
Conditional Variational Autoencoder for Prediction and Feature Recovery Applied to Intrusion Detection in IoT.用于预测和特征恢复的条件变分自编码器在物联网入侵检测中的应用
Sensors (Basel). 2017 Aug 26;17(9):1967. doi: 10.3390/s17091967.
9
Entropic Out-of-Distribution Detection: Seamless Detection of Unknown Examples.信息论离群检测:未知样本的无缝检测。
IEEE Trans Neural Netw Learn Syst. 2022 Jun;33(6):2350-2364. doi: 10.1109/TNNLS.2021.3112897. Epub 2022 Jun 1.
10
A threshold fuzzy entropy based feature selection for medical database classification.基于阈值模糊熵的医学数据库分类特征选择。
Comput Biol Med. 2013 Dec;43(12):2222-9. doi: 10.1016/j.compbiomed.2013.10.016. Epub 2013 Oct 25.

引用本文的文献

1
Scaling DEPP phylogenetic placement to ultra-large reference trees: a tree-aware ensemble approach.将 DEPP 系统发育定位扩展到超大规模参考树:一种基于树的集成方法。
Bioinformatics. 2024 Jun 3;40(6). doi: 10.1093/bioinformatics/btae361.
2
Deep Classification with Linearity-Enhanced Logits to Softmax Function.通过线性增强对数几率到Softmax函数的深度分类
Entropy (Basel). 2023 Apr 27;25(5):727. doi: 10.3390/e25050727.
3
Co-Training for Visual Object Recognition Based on Self-Supervised Models Using a Cross-Entropy Regularization.

本文引用的文献

1
Scoring diverse cellular morphologies in image-based screens with iterative feedback and machine learning.通过迭代反馈和机器学习在基于图像的筛选中对多种细胞形态进行评分。
Proc Natl Acad Sci U S A. 2009 Feb 10;106(6):1826-31. doi: 10.1073/pnas.0808843106. Epub 2009 Feb 2.
基于使用交叉熵正则化的自监督模型的视觉目标识别协同训练
Entropy (Basel). 2021 Apr 1;23(4):423. doi: 10.3390/e23040423.