Henschel Leonie, Kügler David, Zöllei Lilla, Reuter Martin
German Center for Neurodegenerative Diseases (DZNE), Bonn, Germany.
A.A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Boston, MA, USA.
Imaging Neurosci (Camb). 2024 May 30;2:1-26. doi: 10.1162/imag_a_00180. eCollection 2024 May 1.
A robust, fast, and accurate segmentation of neonatal brain images is highly desired to better understand and detect changes during development and disease, specifically considering the rise in imaging studies for this cohort. Yet, the limited availability of ground truth datasets, lack of standardized acquisition protocols, and wide variations of head positioning in the scanner pose challenges for method development. A few automated image analysis pipelines exist for newborn brain Magnetic Resonance Image (MRI) segmentation, but they often rely on time-consuming non-linear spatial registration procedures and require resampling to a common resolution, subject to loss of information due to interpolation and down-sampling. Without registration and image resampling, variations with respect to head positions and voxel resolutions have to be addressed differently. In deep learning, external augmentations such as rotation, translation, and scaling are traditionally used to artificially expand the representation of spatial variability, which subsequently increases both the training dataset size and robustness. However, these transformations in the image space still require resampling, reducing accuracy specifically in the context of label interpolation. We recently introduced the concept of resolution-independence with the Voxel-size Independent Neural Network framework, VINN. Here, we extend this concept by additionally shifting all rigid-transforms into the network architecture with a four degree of freedom (4-DOF) transform module, enabling resolution-aware internal augmentations (VINNA) for deep learning. In this work, we show that VINNA (i) significantly outperforms state-of-the-art external augmentation approaches, (ii) effectively addresses the head variations present specifically in newborn datasets, and (iii) retains high segmentation accuracy across a range of resolutions (0.5-1.0 mm). Furthermore, the 4-DOF transform module together with internal augmentations is a powerful, general approach to implement spatial augmentation without requiring image or label interpolation. The specific network application to newborns will be made publicly available as VINNA4neonates.
为了更好地理解和检测新生儿在发育和疾病过程中的变化,特别是考虑到针对这一群体的成像研究不断增加,人们迫切需要对新生儿脑图像进行强大、快速且准确的分割。然而,真实数据集的可用性有限、缺乏标准化的采集协议以及扫描仪中头部定位的广泛差异,给方法开发带来了挑战。目前存在一些用于新生儿脑磁共振成像(MRI)分割的自动图像分析管道,但它们通常依赖耗时的非线性空间配准程序,并且需要重新采样到共同分辨率,这会因插值和下采样而导致信息丢失。如果不进行配准和图像重采样,就必须以不同方式处理头部位置和体素分辨率的变化。在深度学习中,传统上使用旋转、平移和缩放等外部增强来人为扩展空间变异性的表示,这随后增加了训练数据集的大小和鲁棒性。然而,图像空间中的这些变换仍然需要重新采样,特别是在标签插值的情况下会降低准确性。我们最近通过体素大小独立神经网络框架VINN引入了分辨率独立性的概念。在此,我们通过使用四自由度(4-DOF)变换模块将所有刚性变换额外转移到网络架构中,扩展了这一概念,从而为深度学习实现分辨率感知内部增强(VINNA)。在这项工作中,我们表明VINNA(i)显著优于当前最先进的外部增强方法,(ii)有效解决了新生儿数据集中特有的头部变化问题,并且(iii)在一系列分辨率(0.5 - 1.0毫米)范围内保持高分割准确性。此外,4-DOF变换模块与内部增强一起是一种强大的通用方法,可用于实现空间增强而无需图像或标签插值。针对新生儿的特定网络应用将作为VINNA4neonates公开提供。