Office of Surveillance and Epidemiology, Center for Drug Evaluation and Research, US Food and Drug Administration, Silver Spring, MD, United States.
Jacobs School of Medicine and Biomedical Sciences, Buffalo, NY, United States.
J Med Internet Res. 2024 Jun 6;26:e50274. doi: 10.2196/50274.
Adverse drug reactions are a common cause of morbidity in health care. The US Food and Drug Administration (FDA) evaluates individual case safety reports of adverse events (AEs) after submission to the FDA Adverse Event Reporting System as part of its surveillance activities. Over the past decade, the FDA has explored the application of artificial intelligence (AI) to evaluate these reports to improve the efficiency and scientific rigor of the process. However, a gap remains between AI algorithm development and deployment. This viewpoint aims to describe the lessons learned from our experience and research needed to address both general issues in case-based reasoning using AI and specific needs for individual case safety report assessment. Beginning with the recognition that the trustworthiness of the AI algorithm is the main determinant of its acceptance by human experts, we apply the Diffusion of Innovations theory to help explain why certain algorithms for evaluating AEs at the FDA were accepted by safety reviewers and others were not. This analysis reveals that the process by which clinicians decide from case reports whether a drug is likely to cause an AE is not well defined beyond general principles. This makes the development of high performing, transparent, and explainable AI algorithms challenging, leading to a lack of trust by the safety reviewers. Even accounting for the introduction of large language models, the pharmacovigilance community needs an improved understanding of causal inference and of the cognitive framework for determining the causal relationship between a drug and an AE. We describe specific future research directions that underpin facilitating implementation and trust in AI for drug safety applications, including improved methods for measuring and controlling of algorithmic uncertainty, computational reproducibility, and clear articulation of a cognitive framework for causal inference in case-based reasoning.
药物不良反应是医疗保健中发病率的一个常见原因。美国食品和药物管理局(FDA)在将不良事件(AE)的个体病例安全报告提交给 FDA 不良事件报告系统后,对其进行评估,作为其监测活动的一部分。在过去的十年中,FDA 一直在探索人工智能(AI)在评估这些报告中的应用,以提高该过程的效率和科学严谨性。然而,人工智能算法的开发和部署之间仍然存在差距。本观点旨在描述从我们的经验和研究中吸取的教训,这些经验和研究既需要解决基于案例推理的人工智能应用中的一般问题,也需要解决个体病例安全报告评估的具体需求。首先认识到人工智能算法的可信度是其被人类专家接受的主要决定因素,我们应用创新扩散理论来帮助解释为什么 FDA 评估 AE 的某些算法被安全审查员接受,而其他算法则没有被接受。这项分析表明,临床医生从病例报告中判断药物是否可能引起 AE 的决策过程并不仅限于一般原则。这使得开发高性能、透明和可解释的人工智能算法具有挑战性,导致安全审查员缺乏信任。即使考虑到大型语言模型的引入,药物警戒界也需要更好地理解因果推断以及确定药物与 AE 之间因果关系的认知框架。我们描述了支持在药物安全应用中实施和信任人工智能的具体未来研究方向,包括改进算法不确定性、计算可重复性的测量和控制方法,以及明确阐明基于案例推理的因果推断的认知框架。