Brown Patrick H, Balbo Andrea, Schuck Peter
Protein Biophysics Resource, National Institute of Biomedical Imaging and Bioengineering, National Institutes of Health, Bethesda, Maryland 20892, USA.
Biomacromolecules. 2007 Jun;8(6):2011-24. doi: 10.1021/bm070193j. Epub 2007 May 24.
Analytical ultracentrifugation has reemerged as a widely used tool for the study of ensembles of biological macromolecules to understand, for example, their size-distribution and interactions in free solution. Such information can be obtained from the mathematical analysis of the concentration and signal gradients across the solution column and their evolution in time generated as a result of the gravitational force. In sedimentation velocity analytical ultracentrifugation, this analysis is frequently conducted using high resolution, diffusion-deconvoluted sedimentation coefficient distributions. They are based on Fredholm integral equations, which are ill-posed unless stabilized by regularization. In many fields, maximum entropy and Tikhonov-Phillips regularization are well-established and powerful approaches that calculate the most parsimonious distribution consistent with the data and prior knowledge, in accordance with Occam's razor. In the implementations available in analytical ultracentrifugation, to date, the basic assumption implied is that all sedimentation coefficients are equally likely and that the information retrieved should be condensed to the least amount possible. Frequently, however, more detailed distributions would be warranted by specific detailed prior knowledge on the macromolecular ensemble under study, such as the expectation of the sample to be monodisperse or paucidisperse or the expectation for the migration to establish a bimodal sedimentation pattern based on Gilbert-Jenkins' theory for the migration of chemically reacting systems. So far, such prior knowledge has remained largely unused in the calculation of the sedimentation coefficient or molecular weight distributions or was only applied as constraints. In the present paper, we examine how prior expectations can be built directly into the computational data analysis, conservatively in a way that honors the complete information of the experimental data, whether or not consistent with the prior expectation. Consistent with analogous results in other fields, we find that the use of available prior knowledge can have a dramatic effect on the resulting molecular weight, sedimentation coefficient, and size-and-shape distributions and can significantly increase both their sensitivity and their resolution. Further, the use of multiple alternative prior information allows us to probe the range of possible interpretations consistent with the data.
分析超速离心法已再度成为一种广泛应用的工具,用于研究生物大分子聚集体,以了解其在自由溶液中的大小分布和相互作用等情况。此类信息可通过对溶液柱中浓度和信号梯度及其在重力作用下随时间的演变进行数学分析来获取。在沉降速度分析超速离心中,这种分析通常使用高分辨率、扩散去卷积沉降系数分布来进行。它们基于弗雷德霍姆积分方程,这些方程是不适定的,除非通过正则化进行稳定处理。在许多领域,最大熵和蒂霍诺夫 - 菲利普斯正则化是成熟且强大的方法,它们根据奥卡姆剃刀原理,计算与数据和先验知识一致的最简洁分布。在分析超速离心法现有的实现中,迄今为止隐含的基本假设是所有沉降系数同等可能,并且检索到的信息应尽可能压缩。然而,通常情况下,基于对所研究的大分子聚集体的特定详细先验知识,更详细的分布是合理的,例如期望样品为单分散或寡分散,或者期望根据吉尔伯特 - 詹金斯化学反应体系迁移理论,迁移形成双峰沉降模式。到目前为止,此类先验知识在沉降系数或分子量分布的计算中基本未被使用,或者仅作为约束条件应用。在本文中,我们研究了如何将先验期望直接纳入计算数据分析中,以一种保守的方式尊重实验数据的完整信息,无论其是否与先验期望一致。与其他领域的类似结果一致,我们发现使用可用的先验知识会对所得的分子量、沉降系数以及大小和形状分布产生显著影响,并能显著提高它们的灵敏度和分辨率。此外,使用多种替代先验信息使我们能够探究与数据一致的可能解释范围。