Tate Tia, Patlewicz Grace, Shah Imran
Center for Computational Toxicology and Exposure, Office of Research and Development, U.S. Environmental Protection Agency, Research Triangle Park, North Carolina 27709, USA.
Comput Toxicol. 2024 Mar;29:1-14. doi: 10.1016/j.comtox.2024.100301.
Animal toxicity testing is time and resource intensive, making it difficult to keep pace with the number of substances requiring assessment. Machine learning (ML) models that use chemical structure information and high-throughput experimental data can be helpful in predicting potential toxicity . However, much of the toxicity data used to train ML models is biased with an unequal balance of positives and negatives primarily since substances selected for in vivo testing are expected to elicit some toxicity effect. To investigate the impact this bias had on predictive performance, various sampling approaches were used to balance in vivo toxicity data as part of a supervised ML workflow to predict hepatotoxicity outcomes from chemical structure and/or targeted transcriptomic data. From the chronic, subchronic, developmental, multigenerational reproductive, and subacute repeat-dose testing toxicity outcomes with a minimum of 50 positive and 50 negative substances, 18 different study-toxicity outcome combinations were evaluated in up to 7 ML models. These included Artificial Neural Networks, Random Forests, Bernouilli Naïve Bayes, Gradient Boosting, and Support Vector classification algorithms which were compared with a local approach, Generalised Read-Across (GenRA), a similarity-weighted k-Nearest Neighbour (k-NN) method. The mean CV F1 performance for unbalanced data across all classifiers and descriptors for chronic liver effects was 0.735 (0.0395 SD). Mean CV F1 performance dropped to 0.639 (0.073 SD) with over-sampling approaches though the poorer performance of KNN approaches in some cases contributed to the observed decrease (mean CV F1 performance excluding KNN was 0.697 (0.072 SD)). With under-sampling approaches, the mean CV F1 was 0.523 (0.083 SD). For developmental liver effects, the mean CV F1 performance was much lower with 0.089 (0.111 SD) for unbalanced approaches and 0.149 (0.084 SD) for under-sampling. Over-sampling approaches led to an increase in mean CV F1 performance (0.234, (0.107 SD)) for developmental liver toxicity. Model performance was found to be dependent on dataset, model type, balancing approach and feature selection. Accordingly tailoring ML workflows for predicting toxicity should consider class imbalance and rely on simpler classifiers first.
动物毒性测试耗费时间和资源,因此难以跟上需要评估的物质数量的增长速度。利用化学结构信息和高通量实验数据的机器学习(ML)模型有助于预测潜在毒性。然而,用于训练ML模型的许多毒性数据存在偏差,阳性和阴性数据的平衡不均,主要原因是选择用于体内测试的物质预计会引发某种毒性效应。为了研究这种偏差对预测性能的影响,在一个监督式ML工作流程中,使用了各种采样方法来平衡体内毒性数据,以便从化学结构和/或靶向转录组数据预测肝毒性结果。从慢性、亚慢性、发育、多代生殖和亚急性重复剂量测试的毒性结果中,选取至少50种阳性和50种阴性物质,在多达7种ML模型中评估了18种不同的研究-毒性结果组合。这些模型包括人工神经网络、随机森林、伯努利朴素贝叶斯、梯度提升和支持向量分类算法,并与一种局部方法广义类推法(GenRA)、一种相似性加权k近邻(k-NN)方法进行了比较。所有分类器和描述符针对慢性肝脏效应的不平衡数据的平均CV F1性能为0.735(标准差0.0395)。通过过采样方法,平均CV F1性能降至0.639(标准差0.073),不过在某些情况下KNN方法的较差性能导致了观察到的下降(不包括KNN的平均CV F1性能为0.697(标准差0.072))。采用欠采样方法时,平均CV F1为0.523(标准差0.083)。对于发育性肝脏效应,不平衡方法的平均CV F1性能要低得多,为0.089(标准差0.111),欠采样为0.149(标准差0.084)。过采样方法使发育性肝毒性的平均CV F1性能有所提高(0.234,标准差0.107)。发现模型性能取决于数据集、模型类型、平衡方法和特征选择。因此,为预测毒性量身定制ML工作流程时应考虑类别不平衡,并首先依赖更简单的分类器。