Taleb Ali, Leclerc Sarah, Hussein Raabid, Lalande Alain, Bozorg-Grayeli Alexis
ICMUB Laboratory UMR CNRS 6302, University of Burgundy Franche Comte, 21000, Dijon, France.
Oticon Medical, 06220, Vallauris, France.
Eur Arch Otorhinolaryngol. 2024 Jun;281(6):2921-2930. doi: 10.1007/s00405-023-08403-0. Epub 2024 Jan 10.
Patient-to-image registration is a preliminary step required in surgical navigation based on preoperative images. Human intervention and fiducial markers hamper this task as they are time-consuming and introduce potential errors. We aimed to develop a fully automatic 2D registration system for augmented reality in ear surgery.
CT-scans and corresponding oto-endoscopic videos were collected from 41 patients (58 ears) undergoing ear examination (vestibular schwannoma before surgery, profound hearing loss requiring cochlear implant, suspicion of perilymphatic fistula, contralateral ears in cases of unilateral chronic otitis media). Two to four images were selected from each case. For the training phase, data from patients (75% of the dataset) and 11 cadaveric specimens were used. Tympanic membranes and malleus handles were contoured on both video images and CT-scans by expert surgeons. The algorithm used a U-Net network for detecting the contours of the tympanic membrane and the malleus on both preoperative CT-scans and endoscopic video frames. Then, contours were processed and registered through an iterative closest point algorithm. Validation was performed on 4 cases and testing on 6 cases. Registration error was measured by overlaying both images and measuring the average and Hausdorff distances.
The proposed registration method yielded a precision compatible with ear surgery with a 2D mean overlay error of mm for the incus and mm for the round window. The average Hausdorff distance for these 2 targets was mm and mm respectively. An outlier case with higher errors (2.3 mm and 1.5 mm average Hausdorff distance for incus and round window respectively) was observed in relation to a high discrepancy between the projection angle of the reconstructed CT-scan and the video image. The maximum duration for the overall process was 18 s.
A fully automatic 2D registration method based on a convolutional neural network and applied to ear surgery was developed. The method did not rely on any external fiducial markers nor human intervention for landmark recognition. The method was fast and its precision was compatible with ear surgery.
患者与图像配准是基于术前图像的手术导航所需的初步步骤。人工干预和基准标记会妨碍这项任务,因为它们既耗时又会引入潜在误差。我们旨在开发一种用于耳部手术增强现实的全自动二维配准系统。
收集了41例接受耳部检查患者(58只耳朵)的CT扫描图像和相应的耳内镜视频(手术前的前庭神经鞘瘤、需要人工耳蜗植入的重度听力损失、疑似外淋巴瘘、单侧慢性中耳炎病例的对侧耳朵)。每个病例选取2至4张图像。在训练阶段,使用了患者数据(数据集的75%)和11个尸体标本。专家外科医生在视频图像和CT扫描图像上勾勒出鼓膜和锤骨柄的轮廓。该算法使用U-Net网络在术前CT扫描图像和内镜视频帧上检测鼓膜和锤骨的轮廓。然后,通过迭代最近点算法对轮廓进行处理和配准。对4例进行了验证,对6例进行了测试。通过叠加两幅图像并测量平均距离和豪斯多夫距离来测量配准误差。
所提出的配准方法产生了与耳部手术相匹配的精度,砧骨的二维平均叠加误差为 毫米,圆窗为 毫米。这两个目标的平均豪斯多夫距离分别为 毫米和 毫米。观察到一个误差较高的异常病例(砧骨和圆窗的平均豪斯多夫距离分别为2.3毫米和1.5毫米),这与重建CT扫描的投影角度和视频图像之间的高度差异有关。整个过程的最长持续时间为18秒。
开发了一种基于卷积神经网络并应用于耳部手术的全自动二维配准方法。该方法在识别地标时不依赖任何外部基准标记或人工干预。该方法速度快,其精度与耳部手术相匹配。