Department of Anesthesia and Pain Medicine, The Hospital for Sick Children, Toronto 555 University Avenue, Toronto, ON, M5G 1X8, Canada.
Collaborative Human Immersive Interactive (CHISIL) Laboratory, The Hospital for Sick Children Toronto and Sunnybrook Health Sciences, Toronto, Ontario, Canada.
J Med Syst. 2020 Jan 2;44(2):44. doi: 10.1007/s10916-019-1481-4.
The use of artificial intelligence, including machine learning, is increasing in medicine. Use of machine learning is rising in the prediction of patient outcomes. Machine learning may also be able to enhance and augment anesthesia clinical procedures such as airway management. In this study, we sought to develop a machine learning algorithm that could classify vocal cords and tracheal airway anatomy real-time during video laryngoscopy or bronchoscopy as well as compare the performance of three novel convolutional networks for detecting vocal cords and tracheal rings.
Following institutional approval, a clinical dataset of 775 video laryngoscopy and bronchoscopy videos was used. The dataset was divided into two categories for use for training and testing. We used three convolutional neural networks (CNNs): ResNet, Inception and MobileNet. Backpropagation and a mean squared error loss function were used to assess accuracy as well as minimize bias and variance. Following training, we assessed transferability using the generalization error of the CNN, sensitivity and specificity, average confidence error, outliers, overall confidence percentage, and frames per second for live video feeds. After the training was complete, 22 models using 0 to 25,000 steps were generated and compared.
The overall confidence of classification for the vocal cords and tracheal rings for ResNet, Inception and MobileNet CNNs were as follows: 0.84, 0.78, and 0.64 for vocal cords, respectively, and 0.69, 0.72, 0.54 for tracheal rings, respectively. Transfer learning following additional training resulted in improved accuracy of ResNet and Inception for identifying the vocal cords (with a confidence of 0.96 and 0.93 respectively). The two best performing CNNs, ResNet and Inception, achieved a specificity of 0.985 and 0.971, respectively, and a sensitivity of 0.865 and 0.892, respectively. Inception was able to process the live video feeds at 10 FPS while ResNet processed at 5 FPS. Both were able to pass a feasibility test of identifying vocal cords and tracheal rings in a video feed.
We report the development and evaluation of a CNN that can identify and classify airway anatomy in real time. This neural network demonstrates high performance. The availability of artificial intelligence may improve airway management and bronchoscopy by helping to identify key anatomy real time. Thus, potentially improving performance and outcomes during these procedures. Further, this technology may theoretically be extended to the settings of airway pathology or airway management in the hands of experienced providers. The researchers in this study are exploring the performance of this neural network in clinical trials.
人工智能的应用,包括机器学习,在医学领域的应用越来越多。机器学习在预测患者预后方面的应用也在增加。机器学习还可以增强和补充麻醉临床程序,如气道管理。在这项研究中,我们试图开发一种机器学习算法,该算法可以在视频喉镜或支气管镜检查过程中实时对声带和气管气道解剖结构进行分类,并比较三种新型卷积网络检测声带和气管环的性能。
经机构批准后,使用了 775 个视频喉镜和支气管镜视频的临床数据集。该数据集分为两类,用于训练和测试。我们使用了三种卷积神经网络(CNN):ResNet、Inception 和 MobileNet。使用反向传播和均方误差损失函数来评估准确性,同时最小化偏差和方差。训练完成后,使用 CNN 的泛化误差、灵敏度和特异性、平均置信误差、异常值、总体置信百分比以及实时视频馈送的每秒帧数来评估可转移性。训练完成后,生成并比较了使用 0 到 25,000 步的 22 个模型。
ResNet、Inception 和 MobileNet CNN 对声带和气管环的整体分类置信度如下:声带分别为 0.84、0.78 和 0.64,气管环分别为 0.69、0.72 和 0.54。在进行额外训练后的迁移学习后,ResNet 和 Inception 识别声带的准确性有所提高(置信度分别为 0.96 和 0.93)。表现最好的两个 CNN,ResNet 和 Inception,分别达到了 0.985 和 0.971 的特异性,0.865 和 0.892 的敏感性。Inception 能够以 10 FPS 的速度处理实时视频馈送,而 ResNet 能够以 5 FPS 的速度处理。两者都能够通过识别实时视频馈送中声带和气管环的可行性测试。
我们报告了一种能够实时识别和分类气道解剖结构的 CNN 的开发和评估。这个神经网络表现出了很高的性能。人工智能的出现可能会通过实时帮助识别关键解剖结构来改善气道管理和支气管镜检查,从而潜在地提高这些手术的性能和结果。此外,这项技术理论上可以扩展到气道病理学或有经验的提供者手中的气道管理领域。本研究的研究人员正在临床试验中探索该神经网络的性能。