Science Technology and Society Program, Department of Philosophy and Religious Studies, NC State University, 453 Withers Hall, 101 Lampe Dr, Raleigh, NC, 27607, USA.
Sci Eng Ethics. 2020 Oct;26(5):2461-2472. doi: 10.1007/s11948-020-00242-0.
Autonomous vehicles (AVs)-and accidents they are involved in-attest to the urgent need to consider the ethics of artificial intelligence (AI). The question dominating the discussion so far has been whether we want AVs to behave in a 'selfish' or utilitarian manner. Rather than considering modeling self-driving cars on a single moral system like utilitarianism, one possible way to approach programming for AI would be to reflect recent work in neuroethics. The agent-deed-consequence (ADC) model (Dubljević and Racine in AJOB Neurosci 5(4):3-20, 2014a, Behav Brain Sci 37(5):487-488, 2014b) provides a promising descriptive and normative account while also lending itself well to implementation in AI. The ADC model explains moral judgments by breaking them down into positive or negative intuitive evaluations of the agent, deed, and consequence in any given situation. These intuitive evaluations combine to produce a positive or negative judgment of moral acceptability. For example, the overall judgment of moral acceptability in a situation in which someone committed a deed that is judged as negative (e.g., breaking a law) would be mitigated if the agent had good intentions and the action had a good consequence. This explains the considerable flexibility and stability of human moral judgment that has yet to be replicated in AI. This paper examines the advantages and disadvantages of implementing the ADC model and how the model could inform future work on ethics of AI in general.
自动驾驶汽车(AVs)及其所涉及的事故证明了迫切需要考虑人工智能(AI)的伦理问题。到目前为止,主导讨论的问题是我们是否希望 AVs 以“自私”或功利主义的方式行事。考虑到在单一道德体系(如功利主义)上对自动驾驶汽车进行建模,为 AI 编程的一种可能方法是借鉴神经伦理学的最新研究成果。代理-行为-后果(ADC)模型(Dubljević 和 Racine 在 AJOB Neurosci 5(4):3-20, 2014a,Behav Brain Sci 37(5):487-488, 2014b)提供了一个有前途的描述性和规范性解释,同时也非常适合在 AI 中实现。ADC 模型通过将任何给定情况下的代理、行为和后果的积极或消极直觉评估分解,来解释道德判断。这些直觉评估结合起来产生对道德可接受性的积极或消极判断。例如,在某人做出被判断为负面的行为(例如,违反法律)的情况下,总体道德可接受性的判断将得到缓解,如果代理人有良好的意图并且行为有良好的后果。这解释了人类道德判断的相当大的灵活性和稳定性,这在 AI 中尚未得到复制。本文考察了实施 ADC 模型的优缺点,以及该模型如何为一般人工智能伦理的未来工作提供信息。