Morris Robert R, Kouddous Kareem, Kshirsagar Rohan, Schueller Stephen M
Koko, New York, NY, United States.
Department of Computer Science, Columbia University, New York, NY, United States.
J Med Internet Res. 2018 Jun 26;20(6):e10148. doi: 10.2196/10148.
Conversational agents cannot yet express empathy in nuanced ways that account for the unique circumstances of the user. Agents that possess this faculty could be used to enhance digital mental health interventions.
We sought to design a conversational agent that could express empathic support in ways that might approach, or even match, human capabilities. Another aim was to assess how users might appraise such a system.
Our system used a corpus-based approach to simulate expressed empathy. Responses from an existing pool of online peer support data were repurposed by the agent and presented to the user. Information retrieval techniques and word embeddings were used to select historical responses that best matched a user's concerns. We collected ratings from 37,169 users to evaluate the system. Additionally, we conducted a controlled experiment (N=1284) to test whether the alleged source of a response (human or machine) might change user perceptions.
The majority of responses created by the agent (2986/3770, 79.20%) were deemed acceptable by users. However, users significantly preferred the efforts of their peers (P<.001). This effect was maintained in a controlled study (P=.02), even when the only difference in responses was whether they were framed as coming from a human or a machine.
Our system illustrates a novel way for machines to construct nuanced and personalized empathic utterances. However, the design had significant limitations and further research is needed to make this approach viable. Our controlled study suggests that even in ideal conditions, nonhuman agents may struggle to express empathy as well as humans. The ethical implications of empathic agents, as well as their potential iatrogenic effects, are also discussed.
对话代理尚无法以考虑用户独特情况的细微方式表达同理心。具备这种能力的代理可用于增强数字心理健康干预措施。
我们试图设计一种能够以接近甚至匹配人类能力的方式表达共情支持的对话代理。另一个目标是评估用户如何评价这样一个系统。
我们的系统采用基于语料库的方法来模拟表达的同理心。代理重新利用现有在线同伴支持数据池中的回复并呈现给用户。信息检索技术和词嵌入被用于选择与用户担忧最匹配的历史回复。我们收集了37169名用户的评分以评估该系统。此外,我们进行了一项对照实验(N = 1284),以测试回复的所谓来源(人类或机器)是否可能改变用户的看法。
代理生成的大多数回复(2986/3770,79.20%)被用户认为是可接受的。然而,用户明显更喜欢他们同伴的回复(P <.001)。在对照研究中这种效应依然存在(P =.02),即使回复中唯一的区别在于它们被设定为来自人类还是机器。
我们的系统展示了一种让机器构建细微且个性化共情话语的新方法。然而,该设计存在重大局限性,需要进一步研究以使这种方法可行。我们的对照研究表明,即使在理想条件下,非人类代理在表达同理心方面可能也难以与人类相媲美。我们还讨论了共情代理的伦理含义及其潜在的医源性影响。