Zeng Zhichen, Zhu Ruike, Xia Yinglong, Zeng Hanqing, Tong Hanghang
Department of Computer Science, University of Illinois at Urbana-Champaign, Urbana, IL, USA.
Meta, CA, USA.
Proc Mach Learn Res. 2023;1:40749-40769.
Dictionary learning, which approximates data samples by a set of shared atoms, is a fundamental task in representation learning. However, dictionary learning over graphs, namely graph dictionary learning (GDL), is much more challenging than vectorial data as graphs lie in disparate metric spaces. The sparse literature on GDL formulates the problem from the reconstructive view and often learns linear graph embeddings with a high computational cost. In this paper, we propose a Fused Gromov-Wasserstein (FGW) Mixture Model named FraMe to address the GDL problem from the generative view. Equipped with the graph generation function based on the radial basis function kernel and FGW distance, FraMe generates nonlinear embedding spaces, which, as we theoretically proved, provide a good approximation of the original graph spaces. A fast solution is further proposed on top of the expectation-maximization algorithm with guaranteed convergence. Extensive experiments demonstrate the effectiveness of the obtained node and graph embeddings, and our algorithm achieves significant improvements over the state-of-the-art methods.
字典学习通过一组共享原子来逼近数据样本,是表示学习中的一项基本任务。然而,基于图的字典学习,即图字典学习(GDL),比向量数据更具挑战性,因为图存在于不同的度量空间中。关于GDL的稀疏文献从重构角度阐述问题,并且常常以高计算成本学习线性图嵌入。在本文中,我们提出一种名为FraMe的融合Gromov-Wasserstein(FGW)混合模型,从生成角度解决GDL问题。基于径向基函数核和FGW距离配备图生成函数,FraMe生成非线性嵌入空间,正如我们理论证明的,该空间能很好地逼近原始图空间。在期望最大化算法之上还进一步提出了一种具有收敛保证的快速解决方案。大量实验证明了所获得的节点和图嵌入的有效性,并且我们的算法相对于现有方法有显著改进。