• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

小数据条件下基于噪声时间序列的熵学习

On Entropic Learning from Noisy Time Series in the Small Data Regime.

作者信息

Bassetti Davide, Pospíšil Lukáš, Horenko Illia

机构信息

Faculty of Mathematics, RPTU Kaiserslautern-Landau, Gottlieb-Daimler-Str. 48, 67663 Kaiserslautern, Germany.

Department of Mathematics, Faculty of Civil Engineering, VŠB-TUO, Ludvika Podeste 1875/17, 708 33 Ostrava, Czech Republic.

出版信息

Entropy (Basel). 2024 Jun 28;26(7):553. doi: 10.3390/e26070553.

DOI:10.3390/e26070553
PMID:39056915
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11276242/
Abstract

In this work, we present a novel methodology for performing the supervised classification of time-ordered noisy data; we call this methodology Entropic Sparse Probabilistic Approximation with Markov regularization (eSPA-Markov). It is an extension of entropic learning methodologies, allowing the simultaneous learning of segmentation patterns, entropy-optimal feature space discretizations, and Bayesian classification rules. We prove the conditions for the existence and uniqueness of the learning problem solution and propose a one-shot numerical learning algorithm that-in the leading order-scales linearly in dimension. We show how this technique can be used for the computationally scalable identification of persistent (metastable) regime affiliations and regime switches from high-dimensional non-stationary and noisy time series, i.e., when the size of the data statistics is small compared to their dimensionality and when the noise variance is larger than the variance in the signal. We demonstrate its performance on a set of toy learning problems, comparing eSPA-Markov to state-of-the-art techniques, including deep learning and random forests. We show how this technique can be used for the analysis of noisy time series from DNA and RNA Nanopore sequencing.

摘要

在这项工作中,我们提出了一种用于对时间序列噪声数据进行监督分类的新方法;我们将这种方法称为带马尔可夫正则化的熵稀疏概率近似(eSPA - Markov)。它是熵学习方法的扩展,允许同时学习分割模式、熵最优特征空间离散化和贝叶斯分类规则。我们证明了学习问题解的存在性和唯一性条件,并提出了一种一次性数值学习算法,该算法在主导阶上与维度呈线性比例关系。我们展示了如何将该技术用于从高维非平稳噪声时间序列中以计算可扩展的方式识别持久(亚稳态)状态归属和状态转换,即当数据统计量的大小与其维度相比很小时,以及当噪声方差大于信号方差时。我们在一组玩具学习问题上展示了它的性能,将eSPA - Markov与包括深度学习和随机森林在内的最新技术进行了比较。我们展示了如何将该技术用于分析来自DNA和RNA纳米孔测序的噪声时间序列。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/12df3d50385e/entropy-26-00553-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/b99fef222dc6/entropy-26-00553-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/22e342fc76ae/entropy-26-00553-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/26cac7f2f965/entropy-26-00553-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/576382defdd1/entropy-26-00553-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/d2597627e616/entropy-26-00553-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/41a55998590a/entropy-26-00553-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/a3c5cbbbe7c8/entropy-26-00553-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/12df3d50385e/entropy-26-00553-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/b99fef222dc6/entropy-26-00553-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/22e342fc76ae/entropy-26-00553-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/26cac7f2f965/entropy-26-00553-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/576382defdd1/entropy-26-00553-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/d2597627e616/entropy-26-00553-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/41a55998590a/entropy-26-00553-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/a3c5cbbbe7c8/entropy-26-00553-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/bbc8/11276242/12df3d50385e/entropy-26-00553-g008.jpg

相似文献

1
On Entropic Learning from Noisy Time Series in the Small Data Regime.小数据条件下基于噪声时间序列的熵学习
Entropy (Basel). 2024 Jun 28;26(7):553. doi: 10.3390/e26070553.
2
eSPA+: Scalable Entropy-Optimal Machine Learning Classification for Small Data Problems.eSPA+:针对小数据问题的可扩展熵最优机器学习分类方法
Neural Comput. 2022 Apr 15;34(5):1220-1255. doi: 10.1162/neco_a_01490.
3
On a Scalable Entropic Breaching of the Overfitting Barrier for Small Data Problems in Machine Learning.基于机器学习中小数据问题的可扩展信息泄露突破过拟合障碍
Neural Comput. 2020 Aug;32(8):1563-1579. doi: 10.1162/neco_a_01296. Epub 2020 Jun 10.
4
Large-Scale Sparse Learning From Noisy Tags for Semantic Segmentation.基于含噪标签的大规模稀疏学习的语义分割。
IEEE Trans Cybern. 2018 Jan;48(1):253-263. doi: 10.1109/TCYB.2016.2631528. Epub 2016 Dec 2.
5
Gauge-Optimal Approximate Learning for Small Data Classification.用于小数据分类的尺度最优近似学习
Neural Comput. 2024 May 10;36(6):1198-1227. doi: 10.1162/neco_a_01664.
6
S-CUDA: Self-cleansing unsupervised domain adaptation for medical image segmentation.S-CUDA:用于医学图像分割的自清洁无监督域适应
Med Image Anal. 2021 Dec;74:102214. doi: 10.1016/j.media.2021.102214. Epub 2021 Aug 12.
7
Cheap robust learning of data anomalies with analytically solvable entropic outlier sparsification.通过可解析求解的熵离群值稀疏化实现数据异常的廉价稳健学习。
Proc Natl Acad Sci U S A. 2022 Mar 1;119(9). doi: 10.1073/pnas.2119659119.
8
A hybrid random field model for scalable statistical learning.一种用于可扩展统计学习的混合随机场模型。
Neural Netw. 2009 Jul-Aug;22(5-6):603-13. doi: 10.1016/j.neunet.2009.06.017. Epub 2009 Jul 1.
9
On cheap entropy-sparsified regression learning.关于廉价的熵稀疏回归学习。
Proc Natl Acad Sci U S A. 2023 Jan 3;120(1):e2214972120. doi: 10.1073/pnas.2214972120. Epub 2022 Dec 29.
10
Chained Deep Learning Using Generalized Cross-Entropy for Multiple Annotators Classification.链式深度学习使用广义交叉熵进行多标注分类。
Sensors (Basel). 2023 Mar 28;23(7):3518. doi: 10.3390/s23073518.

本文引用的文献

1
Gauge-Optimal Approximate Learning for Small Data Classification.用于小数据分类的尺度最优近似学习
Neural Comput. 2024 May 10;36(6):1198-1227. doi: 10.1162/neco_a_01664.
2
On cheap entropy-sparsified regression learning.关于廉价的熵稀疏回归学习。
Proc Natl Acad Sci U S A. 2023 Jan 3;120(1):e2214972120. doi: 10.1073/pnas.2214972120. Epub 2022 Dec 29.
3
Low-Cost Probabilistic 3D Denoising with Applications for Ultra-Low-Radiation Computed Tomography.用于超低辐射计算机断层扫描的低成本概率三维去噪
J Imaging. 2022 May 31;8(6):156. doi: 10.3390/jimaging8060156.
4
Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead.停止为高风险决策解释黑箱机器学习模型,转而使用可解释模型。
Nat Mach Intell. 2019 May;1(5):206-215. doi: 10.1038/s42256-019-0048-x. Epub 2019 May 13.
5
eSPA+: Scalable Entropy-Optimal Machine Learning Classification for Small Data Problems.eSPA+:针对小数据问题的可扩展熵最优机器学习分类方法
Neural Comput. 2022 Apr 15;34(5):1220-1255. doi: 10.1162/neco_a_01490.
6
Cheap robust learning of data anomalies with analytically solvable entropic outlier sparsification.通过可解析求解的熵离群值稀疏化实现数据异常的廉价稳健学习。
Proc Natl Acad Sci U S A. 2022 Mar 1;119(9). doi: 10.1073/pnas.2119659119.
7
A roadmap for multi-omics data integration using deep learning.利用深度学习进行多组学数据整合的路线图。
Brief Bioinform. 2022 Jan 17;23(1). doi: 10.1093/bib/bbab454.
8
Using machine learning approaches for multi-omics data analysis: A review.使用机器学习方法进行多组学数据分析:综述
Biotechnol Adv. 2021 Jul-Aug;49:107739. doi: 10.1016/j.biotechadv.2021.107739. Epub 2021 Mar 29.
9
A deeper look into natural sciences with physics-based and data-driven measures.运用基于物理和数据驱动的方法更深入地探究自然科学。
iScience. 2021 Feb 9;24(3):102171. doi: 10.1016/j.isci.2021.102171. eCollection 2021 Mar 19.
10
On a Scalable Entropic Breaching of the Overfitting Barrier for Small Data Problems in Machine Learning.基于机器学习中小数据问题的可扩展信息泄露突破过拟合障碍
Neural Comput. 2020 Aug;32(8):1563-1579. doi: 10.1162/neco_a_01296. Epub 2020 Jun 10.