• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

最小交叉熵模式分类和聚类分析。

Minimum cross-entropy pattern classification and cluster analysis.

机构信息

SENIOR MEMBER, IEEE, Information Technology Division, Naval Research Laboratory, Washington, DC 20375.

出版信息

IEEE Trans Pattern Anal Mach Intell. 1982 Jan;4(1):11-7. doi: 10.1109/tpami.1982.4767189.

DOI:10.1109/tpami.1982.4767189
PMID:21868997
Abstract

This paper considers the problem of classifying an input vector of measurements by a nearest neighbor rule applied to a fixed set of vectors. The fixed vectors are sometimes called characteristic feature vectors, codewords, cluster centers, models, reproductions, etc. The nearest neighbor rule considered uses a non-Euclidean information-theoretic distortion measure that is not a metric, but that nevertheless leads to a classification method that is optimal in a well-defined sense and is also computationally attractive. Furthermore, the distortion measure results in a simple method of computing cluster centroids. Our approach is based on the minimization of cross-entropy (also called discrimination information, directed divergence, K-L number), and can be viewed as a refinement of a general classification method due to Kullback. The refinement exploits special properties of cross-entropy that hold when the probability densities involved happen to be minimum cross-entropy densities. The approach is a generalization of a recently developed speech coding technique called speech coding by vector quantization.

摘要

本文考虑了通过应用于固定向量集的最近邻规则对输入测量向量进行分类的问题。这些固定向量有时被称为特征向量、码字、聚类中心、模型、复制品等。所考虑的最近邻规则使用的是非欧几里得信息论失真度量,它不是度量,但它导致了一种在明确定义的意义上是最优的分类方法,并且在计算上也很有吸引力。此外,失真度量还导致了一种计算聚类中心的简单方法。我们的方法基于交叉熵(也称为鉴别信息、有向散度、K-L 数)的最小化,可以看作是 Kullback 的一种通用分类方法的改进。这种改进利用了交叉熵的特殊性质,当所涉及的概率密度恰好是最小交叉熵密度时,这些性质就成立。这种方法是最近开发的一种语音编码技术的推广,称为矢量量化语音编码。

相似文献

1
Minimum cross-entropy pattern classification and cluster analysis.最小交叉熵模式分类和聚类分析。
IEEE Trans Pattern Anal Mach Intell. 1982 Jan;4(1):11-7. doi: 10.1109/tpami.1982.4767189.
2
Approximate nearest neighbor search by residual vector quantization.基于残差向量量化的近似最近邻搜索。
Sensors (Basel). 2010;10(12):11259-73. doi: 10.3390/s101211259. Epub 2010 Dec 8.
3
Minimum Entropy Rate Simplification of Stochastic Processes.随机过程的最小熵率简化。
IEEE Trans Pattern Anal Mach Intell. 2016 Dec;38(12):2487-2500. doi: 10.1109/TPAMI.2016.2533382. Epub 2016 Feb 23.
4
Successive refinement lattice vector quantization.逐次细化格型矢量量化
IEEE Trans Image Process. 2002;11(12):1337-48. doi: 10.1109/TIP.2002.806235.
5
Prototype Generation Using Self-Organizing Maps for Informativeness-Based Classifier.基于自组织映射的信息性分类器原型生成
Comput Intell Neurosci. 2017;2017:4263064. doi: 10.1155/2017/4263064. Epub 2017 Jul 25.
6
On visualization and aggregation of nearest neighbor classifiers.关于最近邻分类器的可视化与聚合
IEEE Trans Pattern Anal Mach Intell. 2005 Oct;27(10):1592-602. doi: 10.1109/TPAMI.2005.204.
7
Vector quantization for entropy coding of image subbands.图像子带熵编码的矢量量化。
IEEE Trans Image Process. 1992;1(4):526-33. doi: 10.1109/83.199923.
8
Fast nearest neighbor search of entropy-constrained vector quantization.基于约束熵的矢量量化的快速最近邻搜索。
IEEE Trans Image Process. 2000;9(8):1435-7. doi: 10.1109/83.855438.
9
Minimal entropy probability paths between genome families.基因组家族之间的最小熵概率路径。
J Math Biol. 2004 May;48(5):563-90. doi: 10.1007/s00285-003-0248-0. Epub 2003 Dec 2.
10
The Duality of Entropy/Extropy, and Completion of the Kullback Information Complex.熵/负熵的二元性以及库尔贝克信息复合体的完备性。
Entropy (Basel). 2018 Aug 9;20(8):593. doi: 10.3390/e20080593.

引用本文的文献

1
What Loss Functions Do Humans Optimize When They Perform Regression and Classification.人类在进行回归和分类时会优化哪些损失函数。
bioRxiv. 2023 Nov 23:2023.09.19.558376. doi: 10.1101/2023.09.19.558376.
2
Estimating Sentence-like Structure in Synthetic Languages Using Information Topology.利用信息拓扑估计人工语言中类似句子的结构。
Entropy (Basel). 2022 Jun 22;24(7):859. doi: 10.3390/e24070859.
3
An Information Theoretic Approach to Symbolic Learning in Synthetic Languages.一种用于合成语言中符号学习的信息论方法。
Entropy (Basel). 2022 Feb 10;24(2):259. doi: 10.3390/e24020259.
4
The Convex Information Bottleneck Lagrangian.凸信息瓶颈拉格朗日函数。
Entropy (Basel). 2020 Jan 14;22(1):98. doi: 10.3390/e22010098.