Laboratory for Atomistic and Molecular Mechanics, Massachusetts Institute of Technology, 77 Massachusetts Ave., Cambridge, Massachusetts 02139, United States.
Center for Computational Science and Engineering, Schwarzman College of Computing, Massachusetts Institute of Technology, 77 Massachusetts Ave., Cambridge, Massachusetts 02139, United States.
Acc Chem Res. 2022 Dec 6;55(23):3387-3403. doi: 10.1021/acs.accounts.2c00330. Epub 2022 Nov 15.
Humans are continually bombarded with massive amounts of data. To deal with this influx of information, we use the concept of attention in order to perceive the most relevant input from vision, hearing, touch, and others. Thereby, the complex ensemble of signals is used to generate output by querying the processed data in appropriate ways. Attention is also the hallmark of the development of scientific theories, where we elucidate which parts of a problem are critical, often expressed through differential equations. In this Account we review the emergence of attention-based neural networks as a class of approaches that offer many opportunities to describe materials across scales and modalities, including how universal building blocks interact to yield a set of material properties. In fact, the self-assembly of hierarchical, structurally complex, and multifunctional biomaterials remains a grand challenge in modeling, theory, and experiment. Expanding from the process by which material building blocks physically interact to form a type of material, in this Account we view self-assembly as both the functional emergence of properties from interacting building blocks as well as the physical process by which elementary building blocks interact and yield structure and, thereby, functions. This perspective, integrated through the theory of materiomics, allows us to solve multiscale problems with a first-principles-based computational approach based on attention-based neural networks that transform information to feature to property while providing a flexible modeling approach that can integrate theory, simulation, and experiment. Since these models are based on a natural language framework, they offer various benefits including incorporation of general domain knowledge via general-purpose pretraining, which can be accomplished without labeled data or large amounts of lower-quality data. Pretrained models then offer a general-purpose platform that can be fine-tuned to adapt these models to make specific predictions, often with relatively little labeled data. The transferrable power of the language-based modeling approach realizes a neural olog description, where mathematical categorization is learned by multiheaded attention, without domain knowledge in its formulation. It can hence be applied to a range of complex modeling tasks─such as physical field predictions, molecular properties, or structure predictions, all using an identical formulation. This offers a complementary modeling approach that is already finding numerous applications, with great potential to solve complex assembly problems, enabling us to learn, build, and utilize functional categorization of how building blocks yield a range of material functions. In this Account, we demonstrate the approach in various application areas, including protein secondary structure prediction and prediction of normal-mode frequencies as well as predicting mechanical fields near cracks. Unifying these diverse problem areas is the building block approach, where the models are based on a universally applicable platform that offers benefits ranging from transferability, interpretability, and cross-domain pollination of knowledge as exemplified through a transformer model applied to predict how musical compositions infer de novo protein structures. We discuss future potentialities of this approach for a variety of material phenomena across scales, including the use in multiparadigm modeling schemes.
人类不断受到大量数据的冲击。为了应对这种信息涌入,我们使用注意力的概念来感知视觉、听觉、触觉等方面最相关的输入。因此,通过以适当的方式查询处理后的数据,将复杂的信号组合用于生成输出。注意力也是科学理论发展的标志,在科学理论中,我们阐明了问题的哪些部分是关键的,这通常通过微分方程来表达。在本专题介绍中,我们回顾了基于注意力的神经网络的出现,它们是一类提供了许多机会来描述跨尺度和模态的材料的方法,包括通用构建块如何相互作用以产生一组材料性能。事实上,层次结构、结构复杂和多功能生物材料的自组装仍然是建模、理论和实验方面的一大挑战。从物质构建块物理相互作用形成某种材料的过程扩展,在本专题介绍中,我们将自组装视为从相互作用的构建块中产生特性的功能出现,以及基本构建块相互作用并产生结构从而产生功能的物理过程。通过基于注意力的神经网络的材料组学理论进行集成,这种观点允许我们使用基于注意力的神经网络的基于第一性原理的计算方法来解决多尺度问题,该方法将信息转换为特征再转换为特性,同时提供了一种灵活的建模方法,可以整合理论、模拟和实验。由于这些模型基于自然语言框架,因此它们具有各种优势,包括通过通用目的预训练纳入一般领域知识,而无需标记数据或大量低质量数据。经过预训练的模型提供了一个通用平台,可以对其进行微调以适应这些模型,从而进行特定的预测,通常只需相对较少的标记数据。基于语言的建模方法的可转移性实现了神经描述,其中数学分类是通过多头注意力学习的,而在其表述中不需要领域知识。因此,它可以应用于一系列复杂的建模任务,例如物理场预测、分子性质或结构预测,所有这些任务都使用相同的公式。这提供了一种补充的建模方法,已经找到了许多应用,具有解决复杂组装问题的巨大潜力,使我们能够学习、构建和利用构建块产生一系列材料功能的功能分类。在本专题介绍中,我们在各种应用领域展示了该方法,包括蛋白质二级结构预测和预测正常模式频率以及预测裂缝附近的机械场。将这些不同的问题领域统一起来的是构建块方法,其中模型基于普遍适用的平台,具有可转移性、可解释性和跨领域知识授粉等优势,例如通过应用于预测音乐作品如何推断出全新蛋白质结构的转换器模型来说明。我们讨论了这种方法在各种尺度上的各种材料现象中的未来潜力,包括在多范例建模方案中的使用。