Institute of Biomedical Ethics and History of Medicine, University of Zurich, Zurich, Switzerland.
ETH Zurich, Zurich, Switzerland.
Sci Eng Ethics. 2024 Nov 21;30(6):55. doi: 10.1007/s11948-024-00522-z.
We address an open problem in the philosophy of artificial intelligence (AI): how to justify the epistemic attitudes we have towards the trustworthiness of AI systems. The problem is important, as providing reasons to believe that AI systems are worthy of trust is key to appropriately rely on these systems in human-AI interactions. In our approach, we consider the trustworthiness of an AI as a time-relative, composite property of the system with two distinct facets. One is the actual trustworthiness of the AI and the other is the perceived trustworthiness of the system as assessed by its users while interacting with it. We show that credences, namely, beliefs we hold with a degree of confidence, are the appropriate attitude for capturing the facets of the trustworthiness of an AI over time. Then, we introduce a reliabilistic account providing justification to the credences in the trustworthiness of AI, which we derive from Tang's probabilistic theory of justified credence. Our account stipulates that a credence in the trustworthiness of an AI system is justified if and only if it is caused by an assessment process that tends to result in a high proportion of credences for which the actual and perceived trustworthiness of the AI are calibrated. This approach informs research on the ethics of AI and human-AI interactions by providing actionable recommendations on how to measure the reliability of the process through which users perceive the trustworthiness of the system, investigating its calibration to the actual levels of trustworthiness of the AI as well as users' appropriate reliance on the system.
如何为我们对人工智能系统可信度的认知态度提供理由。这个问题很重要,因为为相信人工智能系统值得信任提供理由,是在人机交互中适当依赖这些系统的关键。在我们的方法中,我们将人工智能的可信度视为系统的一个具有两个不同方面的时间相关的复合属性。一方面是人工智能的实际可信度,另一方面是用户在与系统交互时评估的系统的感知可信度。我们表明,信任度,即我们持有一定置信度的信念,是随着时间的推移捕捉人工智能可信度各个方面的适当态度。然后,我们引入了一个可靠主义的解释,为人工智能可信度的信任度提供了理由,这是我们从 Tang 的概率性合理信任度理论中推导出来的。我们的解释规定,如果并且仅当信任度是由一个评估过程引起的,该过程倾向于导致高度的信任度,其中人工智能的实际和感知可信度是校准的,那么对人工智能系统的可信度的信任度就是合理的。这种方法通过提供关于如何通过用户感知系统可信度的过程的可靠性来衡量的可操作建议,为人工智能的伦理和人机交互研究提供了信息,同时也调查了其对人工智能实际可信度水平的校准以及用户对系统的适当依赖程度。