Possolo Michael, Bajcsy Peter
National Institute of Standards and Technology, Gaithersburg, MD 20899, USA.
J Res Natl Inst Stand Technol. 2021 Jun 3;126:126009. doi: 10.6028/jres.126.009. eCollection 2021.
We address the problem of performing exact (tiling-error free) out-of-core semantic segmentation inference of arbitrarily large images using fully convolutional neural networks (FCN). FCN models have the property that once a model is trained, it can be applied on arbitrarily sized images, although it is still constrained by the available GPU memory. This work is motivated by overcoming the GPU memory size constraint without numerically impacting the final result. Our approach is to select a tile size that will fit into GPU memory with a halo border of half the network receptive field. Next, stride across the image by that tile size without the halo. The input tile halos will overlap, while the output tiles join exactly at the seams. Such an approach enables inference to be performed on whole slide microscopy images, such as those generated by a slide scanner. The novelty of this work is in documenting the formulas for determining tile size and stride and then validating them on U-Net and FC-DenseNet architectures. In addition, we quantify the errors due to tiling configurations which do not satisfy the constraints, and we explore the use of architecture effective receptive fields to estimate the tiling parameters.
我们解决了使用全卷积神经网络(FCN)对任意大图像进行精确(无拼接误差)的核外语义分割推理的问题。FCN模型具有这样的特性:一旦模型经过训练,它就可以应用于任意大小的图像,尽管它仍然受到可用GPU内存的限制。这项工作的动机是克服GPU内存大小的限制,同时在数值上不影响最终结果。我们的方法是选择一个能够在GPU内存中容纳且带有网络感受野一半大小光晕边界的切片大小。接下来,以该切片大小在图像上进行步长滑动,不包括光晕部分。输入切片的光晕会重叠,而输出切片在接缝处精确拼接。这种方法能够对整个载玻片显微镜图像进行推理,比如由载玻片扫描仪生成的图像。这项工作的新颖之处在于记录了用于确定切片大小和步长的公式,然后在U-Net和FC-DenseNet架构上对其进行验证。此外,我们对由于不满足约束的拼接配置所导致的误差进行了量化,并探索使用架构有效感受野来估计拼接参数。