Xu Hongteng, Liu Jiachang, Luo Dixin, Carin Lawrence
IEEE Trans Pattern Anal Mach Intell. 2023 Jan;45(1):999-1016. doi: 10.1109/TPAMI.2022.3153126. Epub 2022 Dec 5.
Graph representation is a challenging and significant problem for many real-world applications. In this work, we propose a novel paradigm called "Gromov-Wasserstein Factorization" (GWF) to learn graph representations in a flexible and interpretable way. Given a set of graphs, whose correspondence between nodes is unknown and whose sizes can be different, our GWF model reconstructs each graph by a weighted combination of some "graph factors" under a pseudo-metric called Gromov-Wasserstein (GW) discrepancy. This model leads to a new nonlinear factorization mechanism of the graphs. The graph factors are shared by all the graphs, which represent the typical patterns of the graphs' structures. The weights associated with each graph indicate the graph factors' contributions to the graph's reconstruction, which lead to a permutation-invariant graph representation. We learn the graph factors of the GWF model and the weights of the graphs jointly by minimizing the overall reconstruction error. When learning the model, we reparametrize the graph factors and the weights to unconstrained model parameters and simplify the backpropagation of gradient with the help of the envelope theorem. For the GW discrepancy (the critical training step), we consider two algorithms to compute it, which correspond to the proximal point algorithm (PPA) and Bregman alternating direction method of multipliers (BADMM), respectively. Furthermore, we propose some extensions of the GWF model, including (i) combining with a graph neural network and learning graph representations in an auto-encoding manner, (ii) representing the graphs with node attributes, and (iii) working as a regularizer for semi-supervised graph classification. Experiments on various datasets demonstrate that our GWF model is comparable to the state-of-the-art methods. The graph representations derived by it perform well in graph clustering and classification tasks.
对于许多实际应用而言,图表示是一个具有挑战性且意义重大的问题。在这项工作中,我们提出了一种名为“格罗莫夫 - 瓦瑟斯坦分解”(GWF)的新颖范式,以灵活且可解释的方式学习图表示。给定一组图,其节点之间的对应关系未知且大小可能不同,我们的GWF模型在一种称为格罗莫夫 - 瓦瑟斯坦(GW)差异的伪度量下,通过一些“图因子”的加权组合来重构每个图。该模型导致了一种新的图的非线性分解机制。图因子由所有图共享,代表了图结构的典型模式。与每个图相关联的权重表示图因子对图重构的贡献,从而得到一种置换不变的图表示。我们通过最小化整体重构误差来联合学习GWF模型的图因子和图的权重。在学习模型时,我们将图因子和权重重新参数化为无约束的模型参数,并借助包络定理简化梯度的反向传播。对于GW差异(关键的训练步骤),我们考虑两种算法来计算它,分别对应于近端点算法(PPA)和布雷格曼交替方向乘子法(BADMM)。此外,我们提出了GWF模型的一些扩展,包括(i)与图神经网络相结合并以自动编码方式学习图表示,(ii)用节点属性表示图,以及(iii)作为半监督图分类的正则化器。在各种数据集上的实验表明,我们的GWF模型与当前的先进方法相当。由它导出的图表示在图聚类和分类任务中表现良好。