Linardatos Pantelis, Papastefanopoulos Vasilis, Kotsiantis Sotiris
Department of Mathematics, University of Patras, 26504 Patras, Greece.
Entropy (Basel). 2020 Dec 25;23(1):18. doi: 10.3390/e23010018.
Recent advances in artificial intelligence (AI) have led to its widespread industrial adoption, with machine learning systems demonstrating superhuman performance in a significant number of tasks. However, this surge in performance, has often been achieved through increased model complexity, turning such systems into "black box" approaches and causing uncertainty regarding the way they operate and, ultimately, the way that they come to decisions. This ambiguity has made it problematic for machine learning systems to be adopted in sensitive yet critical domains, where their value could be immense, such as healthcare. As a result, scientific interest in the field of Explainable Artificial Intelligence (XAI), a field that is concerned with the development of new methods that explain and interpret machine learning models, has been tremendously reignited over recent years. This study focuses on machine learning interpretability methods; more specifically, a literature review and taxonomy of these methods are presented, as well as links to their programming implementations, in the hope that this survey would serve as a reference point for both theorists and practitioners.
人工智能(AI)的最新进展已使其在工业中得到广泛应用,机器学习系统在大量任务中展现出超人的性能。然而,这种性能的提升往往是通过增加模型复杂性来实现的,这使得此类系统变成了“黑箱”方法,并导致人们对其运行方式以及最终做出决策的方式存在不确定性。这种模糊性使得机器学习系统难以在敏感但关键的领域(如医疗保健领域,其价值可能巨大)中得到应用。因此,近年来,人们对可解释人工智能(XAI)领域的科学兴趣被极大地重新点燃,该领域关注开发解释和解释机器学习模型的新方法。本研究聚焦于机器学习可解释性方法;更具体地说,本文呈现了这些方法的文献综述和分类法,以及它们的编程实现链接,希望该综述能为理论家和实践者提供一个参考点。