Dept. of Electrical and Computer Engineering, The Johns Hopkins University, Baltimore, MD 21218, USA.
Dept. of Electrical and Computer Engineering, The Johns Hopkins University, Baltimore, MD 21218, USA.
Med Image Anal. 2021 Aug;72:102136. doi: 10.1016/j.media.2021.102136. Epub 2021 Jun 19.
Deep neural networks have been successfully applied to medical image analysis tasks like segmentation and synthesis. However, even if a network is trained on a large dataset from the source domain, its performance on unseen test domains is not guaranteed. The performance drop on data obtained differently from the network's training data is a major problem (known as domain shift) in deploying deep learning in clinical practice. Existing work focuses on retraining the model with data from the test domain, or harmonizing the test domain's data to the network training data. A common practice is to distribute a carefully-trained model to multiple users (e.g., clinical centers), and then each user uses the model to process their own data, which may have a domain shift (e.g., varying imaging parameters and machines). However, the lack of availability of the source training data and the cost of training a new model often prevents the use of known methods to solve user-specific domain shifts. Here, we ask whether we can design a model that, once distributed to users, can quickly adapt itself to each new site without expensive retraining or access to the source training data? In this paper, we propose a model that can adapt based on a single test subject during inference. The model consists of three parts, which are all neural networks: a task model (T) which performs the image analysis task like segmentation; a set of autoencoders (AEs); and a set of adaptors (As). The task model and autoencoders are trained on the source dataset and can be computationally expensive. In the deployment stage, the adaptors are trained to transform the test image and its features to minimize the domain shift as measured by the autoencoders' reconstruction loss. Only the adaptors are optimized during the testing stage with a single test subject thus is computationally efficient. The method was validated on both retinal optical coherence tomography (OCT) image segmentation and magnetic resonance imaging (MRI) T1-weighted to T2-weighted image synthesis. Our method, with its short optimization time for the adaptors (10 iterations on a single test subject) and its additional required disk space for the autoencoders (around 15 MB), can achieve significant performance improvement. Our code is publicly available at: https://github.com/YufanHe/self-domain-adapted-network.
深度神经网络已成功应用于医学图像分析任务,如分割和合成。然而,即使网络是在源域的大型数据集上进行训练的,其在未见测试域上的性能也无法保证。在网络训练数据之外的数据上性能下降是将深度学习应用于临床实践中的一个主要问题(称为域转移)。现有的工作侧重于使用测试域中的数据重新训练模型,或者使测试域的数据与网络训练数据协调一致。一种常见的做法是将精心训练的模型分发给多个用户(例如临床中心),然后每个用户使用模型处理自己的数据,这些数据可能存在域转移(例如,成像参数和机器的变化)。但是,源训练数据的可用性和训练新模型的成本常常会阻止使用已知方法来解决特定用户的域转移问题。在这里,我们是否可以设计一种模型,该模型一旦分发给用户,就可以在没有昂贵的重新训练或访问源训练数据的情况下,快速适应每个新站点?在本文中,我们提出了一种可以在推理过程中基于单个测试对象进行自适应的模型。该模型由三部分组成,均为神经网络:执行图像分析任务(如分割)的任务模型(T);一组自动编码器(AEs);和一组适配器(As)。任务模型和自动编码器是在源数据集上进行训练的,这可能需要大量的计算资源。在部署阶段,适配器被训练以最小化自动编码器的重构损失来转换测试图像及其特征,从而达到域转移。仅在测试阶段使用单个测试对象优化适配器,因此计算效率高。该方法在视网膜光学相干断层扫描(OCT)图像分割和磁共振成像(MRI)T1 加权到 T2 加权图像合成上均得到了验证。我们的方法,其适配器的优化时间很短(单个测试对象上的 10 次迭代),并且自动编码器所需的额外磁盘空间(约 15MB),可以实现显著的性能提升。我们的代码可在以下网址获得:https://github.com/YufanHe/self-domain-adapted-network。