Department of Mathematics, Tufts University, Medford, MA 02155.
Mathematics of AI, IBM Research, Yorktown Heights, NY 10598.
Proc Natl Acad Sci U S A. 2021 Jul 13;118(28). doi: 10.1073/pnas.2015851118.
With the advent of machine learning and its overarching pervasiveness it is imperative to devise ways to represent large datasets efficiently while distilling intrinsic features necessary for subsequent analysis. The primary workhorse used in data dimensionality reduction and feature extraction has been the matrix singular value decomposition (SVD), which presupposes that data have been arranged in matrix format. A primary goal in this study is to show that high-dimensional datasets are more compressible when treated as tensors (i.e., multiway arrays) and compressed via tensor-SVDs under the tensor-tensor product constructs and its generalizations. We begin by proving Eckart-Young optimality results for families of tensor-SVDs under two different truncation strategies. Since such optimality properties can be proven in both matrix and tensor-based algebras, a fundamental question arises: Does the tensor construct subsume the matrix construct in terms of representation efficiency? The answer is positive, as proven by showing that a tensor-tensor representation of an equal dimensional spanning space can be superior to its matrix counterpart. We then use these optimality results to investigate how the compressed representation provided by the truncated tensor SVD is related both theoretically and empirically to its two closest tensor-based analogs, the truncated high-order SVD and the truncated tensor-train SVD.
随着机器学习的出现及其广泛应用,我们必须设计出有效的方法来表示大型数据集,同时提取出后续分析所需的内在特征。在数据降维和特征提取中,主要的工作是矩阵奇异值分解(SVD),它假定数据已经以矩阵格式排列。本研究的主要目标是表明,当将高维数据集视为张量(即多向数组)并通过张量-SVD 在张量张量积结构及其推广下进行压缩时,它们更具可压缩性。我们首先证明了两种不同截断策略下张量-SVD 族的 Eckart-Young 最优性结果。由于这种最优性属性可以在矩阵和张量代数中证明,因此会出现一个基本问题:张量结构在表示效率方面是否包含矩阵结构?答案是肯定的,通过证明具有相同维度的张量子空间的张量张量表示可以优于其矩阵对应物来证明这一点。然后,我们使用这些最优性结果从理论和经验上研究了截断张量 SVD 提供的压缩表示与两个最接近的张量基类似物(截断高阶 SVD 和截断张量训练 SVD)之间的关系。