Behler Jörg
Universität Göttingen, Institut für Physikalische Chemie, Theoretische Chemie, Tammannstraße 6, 37077 Göttingen, Germany.
Chem Rev. 2021 Aug 25;121(16):10037-10072. doi: 10.1021/acs.chemrev.0c00868. Epub 2021 Mar 29.
Since their introduction about 25 years ago, machine learning (ML) potentials have become an important tool in the field of atomistic simulations. After the initial decade, in which neural networks were successfully used to construct potentials for rather small molecular systems, the development of high-dimensional neural network potentials (HDNNPs) in 2007 opened the way for the application of ML potentials in simulations of large systems containing thousands of atoms. To date, many other types of ML potentials have been proposed continuously increasing the range of problems that can be studied. In this review, the methodology of the family of HDNNPs including new recent developments will be discussed using a classification scheme into four generations of potentials, which is also applicable to many other types of ML potentials. The first generation is formed by early neural network potentials designed for low-dimensional systems. High-dimensional neural network potentials established the second generation and are based on three key steps: first, the expression of the total energy as a sum of environment-dependent atomic energy contributions; second, the description of the atomic environments by atom-centered symmetry functions as descriptors fulfilling the requirements of rotational, translational, and permutation invariance; and third, the iterative construction of the reference electronic structure data sets by active learning. In third-generation HDNNPs, in addition, long-range interactions are included employing environment-dependent partial charges expressed by atomic neural networks. In fourth-generation HDNNPs, which are just emerging, in addition, nonlocal phenomena such as long-range charge transfer can be included. The applicability and remaining limitations of HDNNPs are discussed along with an outlook at possible future developments.
自大约25年前被引入以来,机器学习(ML)势能已经成为原子模拟领域的一项重要工具。在最初的十年里,神经网络成功地用于构建相当小的分子系统的势能,2007年高维神经网络势能(HDNNP)的发展为ML势能在包含数千个原子的大系统模拟中的应用开辟了道路。迄今为止,人们不断提出许多其他类型的ML势能,从而不断扩大了可研究问题的范围。在这篇综述中,将使用一种分为四代势能的分类方案来讨论HDNNP家族的方法,包括最近的新进展,这种分类方案也适用于许多其他类型的ML势能。第一代由为低维系统设计的早期神经网络势能构成。高维神经网络势能构成了第二代,它基于三个关键步骤:第一,将总能量表示为与环境相关的原子能量贡献之和;第二,通过以原子为中心的对称函数作为满足旋转、平移和置换不变性要求的描述符来描述原子环境;第三,通过主动学习迭代构建参考电子结构数据集。此外,在第三代HDNNP中,采用由原子神经网络表示的与环境相关的部分电荷来包含长程相互作用。在刚刚出现的第四代HDNNP中,此外还可以包含诸如长程电荷转移等非局部现象。本文将讨论HDNNP的适用性和仍然存在的局限性,并展望未来可能的发展。