• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

胶囊回归的 EM 算法。

An EM Algorithm for Capsule Regression.

机构信息

Department of Computer Science and Engineering, University of California, San Diego, La Jolla, CA 92093-0404

出版信息

Neural Comput. 2021 Jan;33(1):194-226. doi: 10.1162/neco_a_01336. Epub 2020 Oct 20.

DOI:10.1162/neco_a_01336
PMID:33080167
Abstract

We investigate a latent variable model for multinomial classification inspired by recent capsule architectures for visual object recognition (Sabour, Frosst, & Hinton, 2017). Capsule architectures use vectors of hidden unit activities to encode the pose of visual objects in an image, and they use the lengths of these vectors to encode the probabilities that objects are present. Probabilities from different capsules can also be propagated through deep multilayer networks to model the part-whole relationships of more complex objects. Notwithstanding the promise of these networks, there still remains much to understand about capsules as primitive computing elements in their own right. In this letter, we study the problem of capsule regression-a higher-dimensional analog of logistic, probit, and softmax regression in which class probabilities are derived from vectors of competing magnitude. To start, we propose a simple capsule architecture for multinomial classification: the architecture has one capsule per class, and each capsule uses a weight matrix to compute the vector of hidden unit activities for patterns it seeks to recognize. Next, we show how to model these hidden unit activities as latent variables, and we use a squashing nonlinearity to convert their magnitudes as vectors into normalized probabilities for multinomial classification. When different capsules compete to recognize the same pattern, the squashing nonlinearity induces nongaussian terms in the posterior distribution over their latent variables. Nevertheless, we show that exact inference remains tractable and use an expectation-maximization procedure to derive least-squares updates for each capsule's weight matrix. We also present experimental results to demonstrate how these ideas work in practice.

摘要

我们研究了一种基于最近视觉对象识别胶囊架构的多项分类潜在变量模型(Sabour、Frosst 和 Hinton,2017)。胶囊架构使用隐藏单元活动向量来编码图像中视觉对象的姿势,并使用这些向量的长度来编码对象存在的概率。来自不同胶囊的概率也可以通过深层多层网络传播,以模拟更复杂对象的部分-整体关系。尽管这些网络具有很大的潜力,但作为原始计算元素,胶囊本身仍有很多需要理解的地方。在这封信中,我们研究了胶囊回归的问题,这是逻辑回归、概率回归和 softmax 回归的高维模拟,其中类概率是从竞争大小的向量中得出的。首先,我们提出了一种用于多项分类的简单胶囊架构:该架构为每个类别有一个胶囊,每个胶囊使用权重矩阵来计算其试图识别的模式的隐藏单元活动向量。接下来,我们展示了如何将这些隐藏单元活动表示为潜在变量,并使用挤压非线性将它们的大小作为向量转换为多项分类的归一化概率。当不同的胶囊竞争识别相同的模式时,挤压非线性在它们的潜在变量的后验分布中引入了非高斯项。然而,我们表明精确推理仍然是可行的,并使用期望最大化程序为每个胶囊的权重矩阵推导最小二乘更新。我们还提出了实验结果,以展示这些想法在实践中的工作方式。

相似文献

1
An EM Algorithm for Capsule Regression.胶囊回归的 EM 算法。
Neural Comput. 2021 Jan;33(1):194-226. doi: 10.1162/neco_a_01336. Epub 2020 Oct 20.
2
A model selection algorithm for a posteriori probability estimation with neural networks.一种用于神经网络后验概率估计的模型选择算法。
IEEE Trans Neural Netw. 2005 Jul;16(4):799-809. doi: 10.1109/TNN.2005.849826.
3
Synergy between object recognition and image segmentation using the expectation-maximization algorithm.使用期望最大化算法实现目标识别与图像分割之间的协同作用。
IEEE Trans Pattern Anal Mach Intell. 2009 Aug;31(8):1486-501. doi: 10.1109/TPAMI.2008.158.
4
Capsule networks with non-iterative cluster routing.胶囊网络的非迭代聚类路由。
Neural Netw. 2021 Nov;143:690-697. doi: 10.1016/j.neunet.2021.07.032. Epub 2021 Aug 8.
5
Inference and Learning for Generative Capsule Models.生成式胶囊模型的推理与学习
Neural Comput. 2023 Mar 18;35(4):727-761. doi: 10.1162/neco_a_01564.
6
Improved object recognition using neural networks trained to mimic the brain's statistical properties.利用模仿大脑统计特性的神经网络来提高物体识别能力。
Neural Netw. 2020 Nov;131:103-114. doi: 10.1016/j.neunet.2020.07.013. Epub 2020 Jul 29.
7
Image segmentation using hidden Markov Gauss mixture models.使用隐马尔可夫高斯混合模型的图像分割
IEEE Trans Image Process. 2007 Jul;16(7):1902-11. doi: 10.1109/tip.2007.899612.
8
Deep supervised learning with mixture of neural networks.深度监督学习与神经网络混合。
Artif Intell Med. 2020 Jan;102:101764. doi: 10.1016/j.artmed.2019.101764. Epub 2019 Nov 18.
9
Study of a fast discriminative training algorithm for pattern recognition.一种用于模式识别的快速判别训练算法的研究。
IEEE Trans Neural Netw. 2006 Sep;17(5):1212-21. doi: 10.1109/TNN.2006.875992.
10
Fast ML estimation for the mixture of factor analyzers via an ECM algorithm.通过期望条件最大化(ECM)算法对因子分析器混合模型进行快速极大似然估计。
IEEE Trans Neural Netw. 2008 Nov;19(11):1956-61. doi: 10.1109/TNN.2008.2003467.