Sadeghi Vahid, Sanahmadi Yasaman, Behdad Maryam, Vard Alireza, Sharifi Mohsen, Raeisi Ahmad, Nikkhah Mehdi, Mehridehnavi Alireza
Student Research Committee, School of Advanced Technologies in Medicine, Isfahan University of Medical Sciences, Isfahan, Iran.
Medical Image & Signal Processing Research Center, School of Advanced Technologies in Medicine, Isfahan University of Medical Sciences, Isfahan, Iran.
Data Brief. 2024 Sep 10;57:110927. doi: 10.1016/j.dib.2024.110927. eCollection 2024 Dec.
Wireless capsule endoscopy (WCE) is capable of non-invasively visualizing the small intestine, the most complicated segment of the gastrointestinal tract, to detect different types of abnormalities. However, its main drawback is reviewing the vast number of captured images (more than 50,000 frames). The recorded images are only sometimes clear, and different contaminating agents, such as turbid materials and air bubbles, degrade the visualization quality of the WCE images. This condition could cause serious problems such as reducing mucosal view visualization, prolonging recorded video reviewing time, and increasing the risks of missing pathology. On the other hand, accurately quantifying the amount of turbid fluids and bubbles can indicate potential motility malfunction. To assist in developing computer vision-based techniques, we have constructed the first multicentre publicly available clear and contaminated annotated dataset by precisely segmenting 17,593 capsule endoscopy images from three different databases. In contrast to the existing datasets, our dataset has been annotated at the pixel level, discriminating the clear and contaminated regions and subsequently differentiating bubbles and turbid fluids from normal tissue. To create the dataset, we first selected all of the images (2906 frames) in the reduced mucosal view class covering different levels of contamination and randomly selected 12,237 images from the normal class of the copyright-free CC BY 4.0 licensed small bowel capsule endoscopy (SBCE) images from the Kvasir capsule endoscopy database. To mitigate the possible available bias in the mentioned dataset and to increase the sample size, the number of 2077 and 373 images have been stochastically chosen from the SEE-AI project and CECleanliness datasets respectively for the subsequent annotation. Randomly selected images have been annotated with the aid of ImageJ and ITK-SNAP software under the supervision of an expert SBCE reader with extensive experience in gastroenterology and endoscopy. For each image, two binary and tri-colour ground truth (GT) masks have been created in which each pixel has been indexed into two classes (clear and contaminated) and three classes (bubble, turbid fluids, and normal), respectively. To the best of the author's knowledge, there is no implemented clear and contaminated region segmentation on the capsule endoscopy reading software. Curated multicentre dataset can be utilized to implement applicable segmentation algorithms for identification of clear and contaminated regions and discrimination bubbles, as well as turbid fluids from normal tissue in the small intestine. Since the annotated images belong to three different sources, they provide a diverse representation of the clear and contaminated patterns in the WCE images. This diversity is valuable for training the models that are more robust to variations in data characteristics and can generalize well across different subjects and settings. The inclusion of images from three different centres allows for robust cross-validation opportunities, where computer vision-based models can be trained on one centre's annotated images and evaluated on others.
无线胶囊内镜(WCE)能够非侵入性地可视化小肠,这是胃肠道最复杂的部分,以检测不同类型的异常情况。然而,其主要缺点是要查看大量捕获的图像(超过50,000帧)。记录的图像有时并不清晰,并且不同的污染物,如浑浊物质和气泡,会降低WCE图像的可视化质量。这种情况可能会导致严重问题,如减少黏膜视图的可视化、延长记录视频的查看时间以及增加漏诊病变的风险。另一方面,准确量化浑浊液体和气泡的数量可以指示潜在的运动功能障碍。为了协助开发基于计算机视觉的技术,我们通过精确分割来自三个不同数据库的17,593张胶囊内镜图像,构建了第一个多中心公开可用的清晰和受污染的注释数据集。与现有数据集相比,我们的数据集在像素级别进行了注释,区分了清晰和受污染的区域,并随后将气泡和浑浊液体与正常组织区分开来。为了创建该数据集,我们首先选择了覆盖不同污染水平的减少黏膜视图类别中的所有图像(2906帧),并从Kvasir胶囊内镜数据库中随机选择了12,237张来自无版权CC BY 4.0许可的小肠胶囊内镜(SBCE)图像的正常类别图像。为了减轻上述数据集中可能存在的偏差并增加样本量,分别从SEE - AI项目和CECleanliness数据集中随机选择了2077张和373张图像用于后续注释。随机选择的图像在一位在胃肠病学和内镜检查方面有丰富经验的专家SBCE读者的监督下,借助ImageJ和ITK - SNAP软件进行注释。对于每张图像,创建了两个二进制和三色地面真值(GT)掩码,其中每个像素分别被索引为两类(清晰和受污染)和三类(气泡、浑浊液体和正常)。据作者所知,在胶囊内镜阅读软件上尚未实现清晰和受污染区域的分割。精心策划的多中心数据集可用于实施适用的分割算法,以识别清晰和受污染的区域,并区分气泡以及小肠中的浑浊液体与正常组织。由于注释图像来自三个不同的来源它们提供了WCE图像中清晰和受污染模式的多样化表示。这种多样性对于训练对数据特征变化更具鲁棒性且能在不同受试者和设置中良好泛化的模型很有价值。包含来自三个不同中心的图像允许进行强大的交叉验证机会,其中基于计算机视觉的模型可以在一个中心的注释图像上进行训练,并在其他中心进行评估。