Kim Boah, Ye Jong Chul
IEEE Trans Image Process. 2019 Sep 27. doi: 10.1109/TIP.2019.2941265.
Recent state-of-the-art image segmentation algorithms are mostly based on deep neural networks, thanks to their high performance and fast computation time. However, these methods are usually trained in a supervised manner, which requires large number of high quality ground-truth segmentation masks. On the other hand, classical image segmentation approaches such as level-set methods are formulated in a self-supervised manner by minimizing energy functions such as Mumford-Shah functional, so they are still useful to help generation of segmentation masks without labels. Unfortunately, these algorithms are usually computationally expensive and often have limitation in semantic segmentation. In this paper, we propose a novel loss function based on Mumford-Shah functional that can be used in deep-learning based image segmentation without or with small labeled data. This loss function is based on the observation that the softmax layer of deep neural networks has striking similarity to the characteristic function in the Mumford-Shah functional. We show that the new loss function enables semi-supervised and unsupervised segmentation. In addition, our loss function can be also used as a regularized function to enhance supervised semantic segmentation algorithms. Experimental results on multiple datasets demonstrate the effectiveness of the proposed method.
近期的先进图像分割算法大多基于深度神经网络,这得益于其高性能和快速的计算时间。然而,这些方法通常以监督方式进行训练,这需要大量高质量的地面真值分割掩码。另一方面,经典的图像分割方法,如水平集方法,通过最小化能量函数(如Mumford-Shah泛函)以自监督方式制定,因此它们对于帮助生成无标签的分割掩码仍然很有用。不幸的是,这些算法通常计算成本高昂,并且在语义分割方面往往存在局限性。在本文中,我们提出了一种基于Mumford-Shah泛函的新型损失函数,该函数可用于基于深度学习的图像分割,无论是无标签数据还是少量有标签数据的情况。此损失函数基于这样的观察:深度神经网络的softmax层与Mumford-Shah泛函中的特征函数有显著相似性。我们表明,新的损失函数能够实现半监督和无监督分割。此外,我们的损失函数还可以用作正则化函数来增强监督语义分割算法。在多个数据集上的实验结果证明了所提方法的有效性。