Qu Xiaoda, Fan Xiran, Vemuri Baba C
Department of Statistics, University of Florida, Gainesville, FL 32611 USA.
Visa, San Francisco, CA 94128 USA.
IEEE Trans Inf Theory. 2024 Dec;70(12):8961-8974. doi: 10.1109/tit.2024.3468212. Epub 2024 Sep 25.
Distributional approximation is a fundamental problem in machine learning with numerous applications across all fields of science and engineering and beyond. The key challenge in most approximation methods is the need to tackle the intractable normalization constant present in the candidate distributions used to model the data. This intractability is especially common for distributions of manifold-valued random variables such as rotation matrices, orthogonal matrices etc. In this paper, we focus on the distributional approximation problem in Lie groups since they are frequently encountered in many applications including but not limited to, computer vision, robotics, medical imaging and many more. We present a novel Stein's operator on Lie groups leading to a kernel Stein discrepancy (KSD) which is a normalization-free loss function. We present several theoretical results characterizing the properties of this new KSD on Lie groups and its minimizer namely, the minimum KSD estimator (MKSDE). Properties of MKSDE are presented and proved, including strong consistency, CLT and a closed form of the MKSDE for the von Mises-Fisher, the exponential and the Riemannian normal distributions on . Finally, we present several experimental results depicting advantages of MKSDE over maximum likelihood estimation.
分布近似是机器学习中的一个基本问题,在科学、工程及其他领域有着众多应用。大多数近似方法的关键挑战在于需要处理用于对数据进行建模的候选分布中存在的难以处理的归一化常数。这种难处理性在诸如旋转矩阵、正交矩阵等流形值随机变量的分布中尤为常见。在本文中,我们关注李群中的分布近似问题,因为它们在许多应用中经常出现,包括但不限于计算机视觉、机器人技术、医学成像等等。我们提出了一种在李群上的新型斯坦因算子,它导致了一种无归一化损失函数的核斯坦因差异(KSD)。我们给出了几个理论结果,刻画了这种新的李群上的KSD及其最小化器(即最小KSD估计器(MKSDE))的性质。给出并证明了MKSDE的性质,包括强一致性、中心极限定理以及 上冯·米塞斯 - 费希尔分布、指数分布和黎曼正态分布的MKSDE的闭式。最后,我们给出了几个实验结果,展示了MKSDE相对于最大似然估计的优势。