Suppr超能文献

声呐:大规模关联网络学习。

SOFAR: Large-Scale Association Network Learning.

作者信息

Uematsu Yoshimasa, Fan Yingying, Chen Kun, Lv Jinchi, Lin Wei

机构信息

Yoshimasa Uematsu is Assistant Professor, Department of Economics and Management, Tohoku University, Sendai 980-8576, Japan. Yingying Fan is Dean's Associate Professor in Business Administration, Data Sciences and Operations Department, Marshall School of Business, University of Southern California, Los Angeles, CA 90089. Kun Chen is Associate Professor, Department of Statistics, University of Connecticut, Storrs, CT 06269. Jinchi Lv is Kenneth King Stonier Chair in Business Administration and Professor, Data Sciences and Operations Department, Marshall School of Business, University of Southern California, Los Angeles, CA 90089. Wei Lin is Assistant Professor, School of Mathematical Sciences and Center for Statistical Science, Peking University, Beijing, China 100871.

出版信息

IEEE Trans Inf Theory. 2019 Aug;65(8):4924-4939. doi: 10.1109/tit.2019.2909889. Epub 2019 Apr 11.

Abstract

Many modern big data applications feature large scale in both numbers of responses and predictors. Better statistical efficiency and scientific insights can be enabled by understanding the large-scale response-predictor association network structures via layers of sparse latent factors ranked by importance. Yet sparsity and orthogonality have been two largely incompatible goals. To accommodate both features, in this paper we suggest the method of sparse orthogonal factor regression (SOFAR) via the sparse singular value decomposition with orthogonality constrained optimization to learn the underlying association networks, with broad applications to both unsupervised and supervised learning tasks such as biclustering with sparse singular value decomposition, sparse principal component analysis, sparse factor analysis, and spare vector autoregression analysis. Exploiting the framework of convexity-assisted nonconvex optimization, we derive nonasymptotic error bounds for the suggested procedure characterizing the theoretical advantages. The statistical guarantees are powered by an efficient SOFAR algorithm with convergence property. Both computational and theoretical advantages of our procedure are demonstrated with several simulations and real data examples.

摘要

许多现代大数据应用在响应数量和预测变量方面都具有大规模特征。通过按重要性排序的稀疏潜在因子层来理解大规模响应 - 预测变量关联网络结构,可以实现更高的统计效率和科学见解。然而,稀疏性和正交性一直是两个在很大程度上不相容的目标。为了兼顾这两个特征,在本文中,我们提出了稀疏正交因子回归(SOFAR)方法,通过具有正交性约束优化的稀疏奇异值分解来学习潜在的关联网络,该方法在无监督和监督学习任务中都有广泛应用,如使用稀疏奇异值分解的双聚类、稀疏主成分分析、稀疏因子分析和稀疏向量自回归分析。利用凸性辅助非凸优化框架,我们为所提出的过程推导了非渐近误差界,以表征其理论优势。这些统计保证由具有收敛性质的高效SOFAR算法提供支持。我们通过几个模拟和实际数据示例展示了该过程的计算优势和理论优势。

相似文献

1
SOFAR: Large-Scale Association Network Learning.声呐:大规模关联网络学习。
IEEE Trans Inf Theory. 2019 Aug;65(8):4924-4939. doi: 10.1109/tit.2019.2909889. Epub 2019 Apr 11.
2
Sequential Co-Sparse Factor Regression.序贯协同稀疏因子回归
J Comput Graph Stat. 2017;26(4):814-825. doi: 10.1080/10618600.2017.1340891. Epub 2017 Oct 16.
6
Biclustering via sparse singular value decomposition.基于稀疏奇异值分解的双聚类
Biometrics. 2010 Dec;66(4):1087-95. doi: 10.1111/j.1541-0420.2010.01392.x.
7
Unsupervised Feature Selection With Constrained ℓ₂,₀-Norm and Optimized Graph.基于约束ℓ₂,₀范数和优化图的无监督特征选择
IEEE Trans Neural Netw Learn Syst. 2022 Apr;33(4):1702-1713. doi: 10.1109/TNNLS.2020.3043362. Epub 2022 Apr 4.
8
Sparse Reduced Rank Huber Regression in High Dimensions.高维稀疏降秩Huber回归
J Am Stat Assoc. 2023;118(544):2383-2393. doi: 10.1080/01621459.2022.2050243. Epub 2022 Apr 15.
9
Feature Selection With $\ell_{2,1-2}$ Regularization.基于$\ell_{2,1 - 2}$正则化的特征选择
IEEE Trans Neural Netw Learn Syst. 2018 Oct;29(10):4967-4982. doi: 10.1109/TNNLS.2017.2785403. Epub 2018 Jan 15.

本文引用的文献

1
Bayesian sparse reduced rank multivariate regression.贝叶斯稀疏降秩多元回归
J Multivar Anal. 2017 May;157:14-28. doi: 10.1016/j.jmva.2017.02.007. Epub 2017 Mar 4.
3
Principal Component Analysis With Sparse Fused Loadings.具有稀疏融合载荷的主成分分析
J Comput Graph Stat. 2010;19(4):930-946. doi: 10.1198/jcgs.2010.08127.
4
ADAPTIVE ROBUST VARIABLE SELECTION.自适应鲁棒变量选择
Ann Stat. 2014 Feb 1;42(1):324-351. doi: 10.1214/13-AOS1191.
6
Learning regulatory programs by threshold SVD regression.通过阈值奇异值分解回归学习调控程序。
Proc Natl Acad Sci U S A. 2014 Nov 4;111(44):15675-80. doi: 10.1073/pnas.1417808111. Epub 2014 Oct 20.
9
Data, information, knowledge and principle: back to metabolism in KEGG.数据、信息、知识和原理:回到 KEGG 的代谢途径中。
Nucleic Acids Res. 2014 Jan;42(Database issue):D199-205. doi: 10.1093/nar/gkt1076. Epub 2013 Nov 7.

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验