• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于信息价值方法的马尔可夫链约简

Reduction of Markov Chains Using a Value-of-Information-Based Approach.

作者信息

Sledge Isaac J, Príncipe José C

机构信息

Advanced Signal Processing and Automated Target Recognition Branch, US Naval Surface Warfare Center-Panama City Division, Panama City, FL 32407, USA.

Department of Electrical and Computer Engineering, University of Florida, Gainesville, FL 32611, USA.

出版信息

Entropy (Basel). 2019 Mar 30;21(4):349. doi: 10.3390/e21040349.

DOI:10.3390/e21040349
PMID:33267063
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7514833/
Abstract

In this paper, we propose an approach to obtain reduced-order models of Markov chains. Our approach is composed of two information-theoretic processes. The first is a means of comparing pairs of stationary chains on different state spaces, which is done via the negative, modified Kullback-Leibler divergence defined on a model joint space. Model reduction is achieved by solving a value-of-information criterion with respect to this divergence. Optimizing the criterion leads to a probabilistic partitioning of the states in the high-order Markov chain. A single free parameter that emerges through the optimization process dictates both the partition uncertainty and the number of state groups. We provide a data-driven means of choosing the 'optimal' value of this free parameter, which sidesteps needing to a priori know the number of state groups in an arbitrary chain.

摘要

在本文中,我们提出了一种获取马尔可夫链降阶模型的方法。我们的方法由两个信息论过程组成。第一个过程是一种比较不同状态空间上的平稳链对的方法,这是通过在模型联合空间上定义的负的、修正的库尔贝克-莱布勒散度来完成的。通过求解关于这种散度的信息价值准则来实现模型降阶。对该准则进行优化会导致高阶马尔可夫链中状态的概率划分。在优化过程中出现的单个自由参数决定了划分的不确定性和状态组的数量。我们提供了一种数据驱动的方法来选择这个自由参数的“最优”值,从而避免了事先需要知道任意链中状态组数量的问题。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/8842a3e8702d/entropy-21-00349-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/cca849022248/entropy-21-00349-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/d9ad1ff33e80/entropy-21-00349-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/acbf0808f334/entropy-21-00349-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/786df40f2adb/entropy-21-00349-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/b41bbea66f87/entropy-21-00349-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/a6e8bf14fd0a/entropy-21-00349-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/e1296e09e025/entropy-21-00349-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/0eb519388519/entropy-21-00349-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/8842a3e8702d/entropy-21-00349-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/cca849022248/entropy-21-00349-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/d9ad1ff33e80/entropy-21-00349-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/acbf0808f334/entropy-21-00349-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/786df40f2adb/entropy-21-00349-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/b41bbea66f87/entropy-21-00349-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/a6e8bf14fd0a/entropy-21-00349-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/e1296e09e025/entropy-21-00349-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/0eb519388519/entropy-21-00349-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6fd7/7514833/8842a3e8702d/entropy-21-00349-g009.jpg

相似文献

1
Reduction of Markov Chains Using a Value-of-Information-Based Approach.基于信息价值方法的马尔可夫链约简
Entropy (Basel). 2019 Mar 30;21(4):349. doi: 10.3390/e21040349.
2
Performance of Akaike Information Criterion and Bayesian Information Criterion in Selecting Partition Models and Mixture Models.Akaike 信息准则和贝叶斯信息准则在选择划分模型和混合模型中的性能。
Syst Biol. 2023 May 19;72(1):92-105. doi: 10.1093/sysbio/syac081.
3
Analysing grouping of nucleotides in DNA sequences using lumped processes constructed from Markov chains.使用由马尔可夫链构建的集总过程分析DNA序列中的核苷酸分组。
J Math Biol. 2006 Mar;52(3):343-72. doi: 10.1007/s00285-005-0358-y. Epub 2006 Feb 7.
4
Guided Policy Exploration for Markov Decision Processes Using an Uncertainty-Based Value-of-Information Criterion.基于信息价值不确定性准则的马尔可夫决策过程引导策略探索。
IEEE Trans Neural Netw Learn Syst. 2018 Jun;29(6):2080-2098. doi: 10.1109/TNNLS.2018.2812709.
5
Nonparametric identification and maximum likelihood estimation for hidden Markov models.隐马尔可夫模型的非参数识别与最大似然估计
Biometrika. 2016 Jun;103(2):423-434. doi: 10.1093/biomet/asw001. Epub 2016 Mar 28.
6
Optimistic reinforcement learning by forward Kullback-Leibler divergence optimization.基于前向 Kullback-Leibler 散度优化的乐观强化学习。
Neural Netw. 2022 Aug;152:169-180. doi: 10.1016/j.neunet.2022.04.021. Epub 2022 Apr 21.
7
Kullback-Leibler Divergence-Based Differential Evolution Markov Chain Filter for Global Localization of Mobile Robots.基于库尔贝克-莱布勒散度的差分进化马尔可夫链滤波器用于移动机器人的全局定位
Sensors (Basel). 2015 Sep 16;15(9):23431-58. doi: 10.3390/s150923431.
8
Approximating Markov chains.近似马尔可夫链。
Proc Natl Acad Sci U S A. 1992 May 15;89(10):4432-6. doi: 10.1073/pnas.89.10.4432.
9
Kullback-Leibler Divergence Based Probabilistic Approach for Device-Free Localization Using Channel State Information.基于 Kullback-Leibler 散度的概率方法在使用信道状态信息的无设备定位中的应用。
Sensors (Basel). 2019 Nov 3;19(21):4783. doi: 10.3390/s19214783.
10
Finite state automata resulting from temporal information maximization and a temporal learning rule.
Neural Comput. 2005 Oct;17(10):2258-90. doi: 10.1162/0899766054615671.

本文引用的文献

1
An Analysis of the Value of Information When Exploring Stochastic, Discrete Multi-Armed Bandits.探索随机离散多臂老虎机时信息价值的分析
Entropy (Basel). 2018 Feb 28;20(3):155. doi: 10.3390/e20030155.
2
Guided Policy Exploration for Markov Decision Processes Using an Uncertainty-Based Value-of-Information Criterion.基于信息价值不确定性准则的马尔可夫决策过程引导策略探索。
IEEE Trans Neural Netw Learn Syst. 2018 Jun;29(6):2080-2098. doi: 10.1109/TNNLS.2018.2812709.