Centre for Medical Ethics, Institute of Health and Society, Faculty of Medicine, University of Oslo, Kirkeveien 166, Fredrik Holsts hus, Oslo, 0450, Norway.
Med Health Care Philos. 2023 Dec;26(4):615-623. doi: 10.1007/s11019-023-10175-7. Epub 2023 Aug 29.
The difficulty of explaining the outputs of artificial intelligence (AI) models and what has led to them is a notorious ethical problem wherever these technologies are applied, including in the medical domain, and one that has no obvious solution. This paper examines the proposal, made by Luciano Floridi and colleagues, to include a new 'principle of explicability' alongside the traditional four principles of bioethics that make up the theory of 'principlism'. It specifically responds to a recent set of criticisms that challenge the supposed need for such a principle to perform an enabling role in relation to the traditional four principles and therefore suggest that these four are sufficient without the addition of explicability. The paper challenges the critics' premise that explicability cannot be an ethical principle like the classic four because it is explicitly subordinate to them. It argues instead that principlism in its original formulation locates the justification for ethical principles in a midlevel position such that they mediate between the most general moral norms and the contextual requirements of medicine. This conception of an ethical principle then provides a mold for an approach to explicability on which it functions as an enabling principle that unifies technical/epistemic demands on AI and the requirements of high-level ethical theories. The paper finishes by anticipating an objection that decision-making by clinicians and AI fall equally, but implausibly, under the principle of explicability's scope, which it rejects on the grounds that human decisions, unlike AI's, can be explained by their social environments.
解释人工智能 (AI) 模型的输出以及导致这些输出的原因的难度是这些技术应用中(包括医学领域)一个臭名昭著的道德问题,而且这个问题没有明显的解决方案。本文探讨了 Luciano Floridi 和同事提出的建议,即在构成“原则主义”理论的传统四项生物伦理学原则之外,纳入一项新的“可解释性原则”。它特别回应了最近的一系列批评,这些批评质疑了这样一个原则在传统四项原则方面发挥支持作用的必要性,因此表明,在不需要可解释性的情况下,这四项原则就足够了。本文反驳了批评者的前提,即可解释性不能像经典的四项原则那样成为一项伦理原则,因为它明确从属于它们。相反,它认为,原则主义在其最初的表述中,将伦理原则的理由定位在一个中层位置,以便它们在最一般的道德规范和医学的语境要求之间进行调解。然后,这种伦理原则的概念为一种可解释性方法提供了一个模式,它作为一个支持性原则发挥作用,将对 AI 的技术/知识要求与高级伦理理论的要求统一起来。本文最后预计会出现一个反对意见,即临床医生和 AI 的决策同样但不合理地属于可解释性原则的范围,本文对此表示反对,理由是人类的决策与 AI 的决策不同,可以通过其社会环境来解释。