Department of Radiology, Massachusetts General Hospital (MGH) and Harvard Medical School, Boston, MA, USA.
Department of Radiation Oncology, Yonsei University College of Medicine, Seoul, South Korea.
Nat Commun. 2024 Oct 24;15(1):9186. doi: 10.1038/s41467-024-53387-y.
Target volume contouring for radiation therapy is considered significantly more challenging than the normal organ segmentation tasks as it necessitates the utilization of both image and text-based clinical information. Inspired by the recent advancement of large language models (LLMs) that can facilitate the integration of the textural information and images, here we present an LLM-driven multimodal artificial intelligence (AI), namely LLMSeg, that utilizes the clinical information and is applicable to the challenging task of 3-dimensional context-aware target volume delineation for radiation oncology. We validate our proposed LLMSeg within the context of breast cancer radiotherapy using external validation and data-insufficient environments, which attributes highly conducive to real-world applications. We demonstrate that the proposed multimodal LLMSeg exhibits markedly improved performance compared to conventional unimodal AI models, particularly exhibiting robust generalization performance and data-efficiency.
放射治疗的靶区勾画被认为比正常器官分割任务更具挑战性,因为它需要同时利用图像和基于文本的临床信息。受最近在能够促进纹理信息和图像融合的大型语言模型(LLM)方面的进展的启发,我们提出了一种基于 LLM 的多模态人工智能(AI),即 LLMSeg,它利用临床信息,适用于放射肿瘤学中具有挑战性的三维上下文感知靶区勾画任务。我们在乳腺癌放射治疗的背景下使用外部验证和数据不足的环境来验证我们提出的 LLMSeg,这非常有利于实际应用。我们证明,与传统的单模态 AI 模型相比,所提出的多模态 LLMSeg 表现出明显更好的性能,特别是表现出强大的泛化性能和数据效率。