Google DeepMind.
Department of Psychology, Yale University.
Cogn Sci. 2023 Aug;47(8):e13315. doi: 10.1111/cogs.13315.
In developing artificial intelligence (AI), researchers often benchmark against human performance as a measure of progress. Is this kind of comparison possible for moral cognition? Given that human moral judgment often hinges on intangible properties like "intention" which may have no natural analog in artificial agents, it may prove difficult to design a "like-for-like" comparison between the moral behavior of artificial and human agents. What would a measure of moral behavior for both humans and AI look like? We unravel the complexity of this question by discussing examples within reinforcement learning and generative AI, and we examine how the puzzle of evaluating artificial agents' moral cognition remains open for further investigation within cognitive science.
在开发人工智能 (AI) 时,研究人员通常将其与人类表现进行基准测试,作为衡量进展的一种手段。这种比较在道德认知方面是否可行?鉴于人类的道德判断往往取决于无形的属性,如“意图”,而这些属性在人工智能代理中可能没有自然的对应物,因此可能很难在人工智能代理和人类代理的道德行为之间设计出“一对一”的比较。人类和人工智能的道德行为衡量标准会是什么样的?我们通过讨论强化学习和生成人工智能中的例子来解开这个问题的复杂性,我们还研究了如何在认知科学中进一步研究评估人工智能代理的道德认知的难题。