García-Dorado Aurora, Gallego Araceli
Departamento de Genética, Facultad de Biología, Universidad Complutense de Madrid, Spain.
Genetics. 2003 Jun;164(2):807-19. doi: 10.1093/genetics/164.2.807.
We simulated single-generation data for a fitness trait in mutation-accumulation (MA) experiments, and we compared three methods of analysis. Bateman-Mukai (BM) and maximum likelihood (ML) need information on both the MA lines and control lines, while minimum distance (MD) can be applied with or without the control. Both MD and ML assume gamma-distributed mutational effects. ML estimates of the rate of deleterious mutation had larger mean square error (MSE) than MD or BM had due to large outliers. MD estimates obtained by ignoring the mean decline observed from comparison to a control are often better than those obtained using that information. When effects are simulated using the gamma distribution, reducing the precision with which the trait is assayed increases the probability of obtaining no ML or MD estimates but causes no appreciable increase of the MSE. When the residual errors for the means of the simulated lines are sampled from the empirical distribution in a MA experiment, instead of from a normal one, the MSEs of BM, ML, and MD are practically unaffected. When the simulated gamma distribution accounts for a high rate of mild deleterious mutation, BM detects only approximately 30% of the true deleterious mutation rate, while MD or ML detects substantially larger fractions. To test the robustness of the methods, we also added a high rate of common contaminant mutations with constant mild deleterious effect to a low rate of mutations with gamma-distributed deleterious effects and moderate average. In that case, BM detects roughly the same fraction as before, regardless of the precision of the assay, while ML fails to provide estimates. However, MD estimates are obtained by ignoring the control information, detecting approximately 70% of the total mutation rate when the mean of the lines is assayed with good precision, but only 15% for low-precision assays. Contaminant mutations with only tiny deleterious effects could not be detected with acceptable accuracy by any of the above methods.
我们在突变积累(MA)实验中模拟了一个适合度性状的单代数据,并比较了三种分析方法。贝特曼 - 向井(BM)法和最大似然(ML)法需要MA品系和对照品系的信息,而最小距离(MD)法无论有无对照均可应用。MD法和ML法都假定突变效应呈伽马分布。由于存在较大的异常值,有害突变率的ML估计值的均方误差(MSE)比MD法或BM法的要大。忽略与对照比较时观察到的平均下降而获得的MD估计值通常比使用该信息获得的估计值更好。当使用伽马分布模拟效应时,降低性状测定的精度会增加无法获得ML或MD估计值的概率,但不会导致MSE有明显增加。当模拟品系均值的残差误差从MA实验的经验分布中抽样,而不是从正态分布中抽样时,BM法、ML法和MD法的MSE实际上不受影响。当模拟的伽马分布考虑到高比例的轻度有害突变时,BM法仅能检测到约30%的真实有害突变率,而MD法或ML法能检测到的比例要大得多。为了测试这些方法的稳健性,我们还将具有恒定轻度有害效应的高比例常见污染突变添加到具有伽马分布有害效应且平均效应适中的低比例突变中。在这种情况下,无论测定精度如何,BM法检测到的比例与之前大致相同,而ML法无法提供估计值。然而,MD估计值是通过忽略对照信息获得的,当品系均值测定精度良好时,能检测到约70%的总突变率,但在低精度测定时仅为15%。上述任何方法都无法以可接受的精度检测到只有微小有害效应的污染突变。