Kayode Gbolade O, Montemore Matthew M
Department of Chemical and Biomolecular Engineering, Tulane University, New Orleans, Louisiana 70118, United States.
JACS Au. 2023 Dec 19;4(1):80-91. doi: 10.1021/jacsau.3c00419. eCollection 2024 Jan 22.
Machine learning has been successfully applied in recent years to screen materials for a variety of applications. However, despite recent advances, most screening-based machine learning approaches are limited in generality and transferability, requiring new models to be created from scratch for each new application. This is particularly apparent in catalysis, where there are many possible intermediates and transition states of interest in addition to a large number of potential catalytic materials. In this work, we developed a new machine learning framework that is built on chemical principles and allows the creation of general, interpretable, reusable models. Our new architecture uses latent variables to create a set of submodels that each take on a relatively simple learning task, leading to higher data efficiency and promoting transfer learning. This architecture infuses fundamental chemical principles, such as the existence of elements as discrete entities. We show that this architecture allows for the creation of models that can be reused for many different applications, providing significant improvements in efficiency and convenience. For example, our architecture allows simultaneous prediction of adsorption energies for many adsorbates on a broad array of alloy surfaces with mean absolute errors (MAEs) around 0.20-0.25 eV. The integration of latent variables provides physical interpretability, as predictions can be explained in terms of the learned chemical environment as represented by the latent space. Further, these latent variables also serve as new feature representations, allowing efficient transfer learning. For example, new models with useful levels of accuracy can be created with less than 10 data points, including transfer learning to an experimental data set with an MAE less than 0.15 eV. Lastly, we show that our new machine learning architecture is general and robust enough to handle heterogeneous and multifidelity data sets, allowing researchers to leverage existing data sets to speed up screening using their own computational setup.
近年来,机器学习已成功应用于筛选适用于各种应用的材料。然而,尽管有最近的进展,但大多数基于筛选的机器学习方法在通用性和可转移性方面都受到限制,需要为每个新应用从头创建新模型。这在催化领域尤为明显,除了大量潜在的催化材料外,还有许多可能感兴趣的中间体和过渡态。在这项工作中,我们开发了一种基于化学原理的新机器学习框架,该框架允许创建通用、可解释、可重复使用的模型。我们的新架构使用潜在变量来创建一组子模型,每个子模型承担相对简单的学习任务,从而提高数据效率并促进迁移学习。这种架构融入了基本的化学原理,例如元素作为离散实体的存在。我们表明,这种架构允许创建可用于许多不同应用的模型,在效率和便利性方面有显著提高。例如,我们的架构允许同时预测多种吸附质在各种合金表面上的吸附能,平均绝对误差(MAE)约为0.20 - 0.25电子伏特。潜在变量的整合提供了物理可解释性,因为预测可以根据潜在空间所代表的学习化学环境来解释。此外,这些潜在变量还用作新的特征表示,允许进行有效的迁移学习。例如,可以用少于10个数据点创建具有有用精度水平的新模型,包括迁移学习到MAE小于0.15电子伏特的实验数据集。最后,我们表明我们的新机器学习架构具有足够的通用性和鲁棒性来处理异构和多保真度数据集,使研究人员能够利用现有数据集,使用他们自己的计算设置来加速筛选。