School of Plant Sciences and Food Security, Tel Aviv University, Ramat Aviv, Tel Aviv 69978, Israel.
The Shmunis School of Biomedicine and Cancer Research, Tel Aviv University, Ramat Aviv, Tel Aviv 69978, Israel.
Mol Biol Evol. 2024 Jun 1;41(6). doi: 10.1093/molbev/msae105.
The computational search for the maximum-likelihood phylogenetic tree is an NP-hard problem. As such, current tree search algorithms might result in a tree that is the local optima, not the global one. Here, we introduce a paradigm shift for predicting the maximum-likelihood tree, by approximating long-term gains of likelihood rather than maximizing likelihood gain at each step of the search. Our proposed approach harnesses the power of reinforcement learning to learn an optimal search strategy, aiming at the global optimum of the search space. We show that when analyzing empirical data containing dozens of sequences, the log-likelihood improvement from the starting tree obtained by the reinforcement learning-based agent was 0.969 or higher compared to that achieved by current state-of-the-art techniques. Notably, this performance is attained without the need to perform costly likelihood optimizations apart from the training process, thus potentially allowing for an exponential increase in runtime. We exemplify this for data sets containing 15 sequences of length 18,000 bp and demonstrate that the reinforcement learning-based method is roughly three times faster than the state-of-the-art software. This study illustrates the potential of reinforcement learning in addressing the challenges of phylogenetic tree reconstruction.
计算寻找最大似然系统发育树是一个 NP 难问题。因此,当前的树搜索算法可能会导致得到一个局部最优的树,而不是全局最优的树。在这里,我们引入了一种预测最大似然树的范式转变,通过近似长期似然收益,而不是在搜索的每一步最大化似然收益。我们提出的方法利用强化学习的力量来学习最佳的搜索策略,旨在搜索空间的全局最优。我们表明,当分析包含数十个序列的经验数据时,强化学习代理从起始树获得的对数似然改进比当前最先进技术获得的对数似然改进高 0.969 或更高。值得注意的是,这一性能是在除了训练过程之外不需要进行昂贵的似然优化的情况下实现的,因此潜在地允许运行时间呈指数级增长。我们用包含 15 个长度为 18000bp 的序列的数据集来说明这一点,并表明基于强化学习的方法比最先进的软件快大约三倍。这项研究说明了强化学习在解决系统发育树重建挑战方面的潜力。