Suppr超能文献

多模态深度学习在根尖片自动拼接中的应用。

Multi-modal deep learning for automated assembly of periapical radiographs.

机构信息

Department of Oral Diagnostics, Digital Health and Health Services Research, Charité-Universitätsmedizin Berlin, 14197 Berlin, Germany.

Department of Oral Diagnostics, Digital Health and Health Services Research, Charité-Universitätsmedizin Berlin, 14197 Berlin, Germany; ITU/WHO Focus Group AI4Health, Topic Group Dental Diagnostics and Digital Dentistry, Geneva, Switzerland.

出版信息

J Dent. 2023 Aug;135:104588. doi: 10.1016/j.jdent.2023.104588. Epub 2023 Jun 21.

Abstract

OBJECTIVES

Periapical radiographs are oftentimes taken in series to display all teeth present in the oral cavity. Our aim was to automatically assemble such a series of periapical radiographs into an anatomically correct status using a multi-modal deep learning model.

METHODS

4,707 periapical images from 387 patients (on average, 12 images per patient) were used. Radiographs were labeled according to their field of view and the dataset split into a training, validation, and test set, stratified by patient. In addition to the radiograph the timestamp of image generation was extracted and abstracted as follows: A matrix, containing the normalized timestamps of all images of a patient was constructed, representing the order in which images were taken, providing temporal context information to the deep learning model. Using the image data together with the time sequence data a multi-modal deep learning model consisting of two residual convolutional neural networks (ResNet-152 for image data, ResNet-50 for time data) was trained. Additionally, two uni-modal models were trained on image data and time data, respectively. A custom scoring technique was used to measure model performance.

RESULTS

Multi-modal deep learning outperformed both uni-modal image-based learning (p<0.001) and time-based learning (p<0.05). The multi-modal deep learning model predicted tooth labels with an F1-score, sensitivity and precision of 0.79, respectively, and an accuracy of 0.99. 37 out of 77 patient datasets were fully correctly assembled by multi-modal learning; in the remaining ones, usually only one image was incorrectly labeled.

CONCLUSIONS

Multi-modal modeling allowed automated assembly of periapical radiographs and outperformed both uni-modal models. Dental machine learning models can benefit from additional data modalities.

CLINICAL SIGNIFICANCE

Like humans, deep learning models may profit from multiple data sources for decision-making. We demonstrate how multi-modal learning can assist assembling periapical radiographs into an anatomically correct status. Multi-modal learning should be considered for more complex tasks, as clinically a wealth of data is usually available and could be leveraged.

摘要

目的

根尖片通常是一系列拍摄的,以显示口腔中所有存在的牙齿。我们的目的是使用多模态深度学习模型自动将这样一系列根尖片组装成解剖学上正确的状态。

方法

使用了 387 名患者的 4707 张根尖片(平均每个患者 12 张)。根据视野对射线照片进行标记,并根据患者分层将数据集分为训练集、验证集和测试集。除射线照片外,还提取并抽象了图像生成的时间戳,具体方法是:构建一个包含患者所有图像归一化时间戳的矩阵,代表图像拍摄的顺序,为深度学习模型提供时间上下文信息。使用图像数据和时间序列数据,训练了一个由两个残差卷积神经网络(用于图像数据的 ResNet-152 和用于时间数据的 ResNet-50)组成的多模态深度学习模型。此外,还分别在图像数据和时间数据上训练了两个单模态模型。使用自定义评分技术来衡量模型性能。

结果

多模态深度学习优于单模态基于图像的学习(p<0.001)和基于时间的学习(p<0.05)。多模态深度学习模型预测牙齿标签的 F1 得分为 0.79,灵敏度和精度分别为 0.79 和 0.99。多模态学习正确组装了 77 个患者数据集的 37 个;在其余的患者中,通常只有一张图像标记不正确。

结论

多模态建模允许自动组装根尖片,并优于两种单模态模型。牙科机器学习模型可以从其他数据模态中受益。

临床意义

像人类一样,深度学习模型可能会从多个数据源中受益,以做出决策。我们展示了多模态学习如何帮助将根尖片组装成解剖学上正确的状态。对于更复杂的任务,应考虑多模态学习,因为临床上通常有大量可用的数据,可以加以利用。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验