Hakim Joe B, Painter Jeffery L, Ramcharran Darmendra, Kara Vijay, Powell Greg, Sobczak Paulina, Sato Chiho, Bate Andrew, Beam Andrew
Harvard-MIT Department of Health Sciences and Technology, Cambridge, MA, USA.
GSK, Durham, NC, USA.
Sci Rep. 2025 Jul 31;15(1):27886. doi: 10.1038/s41598-025-09138-0.
Large language models (LLMs) are useful tools with the capacity for performing specific types of knowledge work at an effective scale. However, LLM deployments in high-risk and safety-critical domains pose unique challenges, notably the issue of "hallucinations", where LLMs can generate fabricated information. This is particularly concerning in settings such as drug safety, where inaccuracies could lead to patient harm. To mitigate these risks, we have developed and demonstrated a proof of concept suite of guardrails specifically designed to mitigate certain types of hallucinations and errors for drug safety, with potential applicability to other medical safety-critical contexts. These guardrails include mechanisms to detect anomalous documents to prevent the ingestion of inappropriate data, identify incorrect drug names or adverse event terms, and convey uncertainty in generated content. We integrated these guardrails with an LLM fine-tuned for a text-to-text task, which involves converting both structured and unstructured data within adverse event reports into natural language. This method was applied to translate individual case safety reports, demonstrating effective application in a pharmacovigilance processing task. Our guardrail framework offers a set of tools with broad applicability across various domains, ensuring LLMs can be safely used in high-risk situations by eliminating the occurrence of key errors, including the generation of incorrect pharmacovigilance-related terms, thus adhering to stringent regulatory and quality standards in medical safety-critical environments.
大语言模型(LLMs)是有用的工具,能够在有效规模上执行特定类型的知识工作。然而,在高风险和安全关键领域部署大语言模型带来了独特的挑战,尤其是“幻觉”问题,即大语言模型可能生成虚假信息。在药物安全等环境中,这一问题尤其令人担忧,因为不准确的信息可能导致患者受到伤害。为了降低这些风险,我们开发并展示了一套概念验证的防护措施,专门设计用于减轻药物安全中某些类型的幻觉和错误,并有可能应用于其他医疗安全关键环境。这些防护措施包括检测异常文档以防止摄入不适当数据的机制、识别错误的药物名称或不良事件术语,以及传达生成内容中的不确定性。我们将这些防护措施与针对文本到文本任务进行微调的大语言模型集成,该任务涉及将不良事件报告中的结构化和非结构化数据转换为自然语言。此方法应用于翻译个体病例安全报告,证明了其在药物警戒处理任务中的有效应用。我们的防护框架提供了一套具有广泛适用性的工具,通过消除关键错误的发生,确保大语言模型能够在高风险情况下安全使用,包括生成与药物警戒相关的错误术语,从而在医疗安全关键环境中遵守严格的监管和质量标准。