Li Ru, Wu Chi-Hao, Liu Shuaicheng, Wang Jue, Wang Guangfu, Liu Guanghui, Zeng Bing
IEEE Trans Image Process. 2021;30:374-385. doi: 10.1109/TIP.2020.3036754. Epub 2020 Nov 23.
The paper proposes a solution to effectively handle salient regions for style transfer between unpaired datasets. Recently, Generative Adversarial Networks (GAN) have demonstrated their potentials of translating images from source domain X to target domain Y in the absence of paired examples. However, such a translation cannot guarantee to generate high perceptual quality results. Existing style transfer methods work well with relatively uniform content, they often fail to capture geometric or structural patterns that always belong to salient regions. Detail losses in structured regions and undesired artifacts in smooth regions are unavoidable even if each individual region is correctly transferred into the target style. In this paper, we propose SDP-GAN, a GAN-based network for solving such problems while generating enjoyable style transfer results. We introduce a saliency network, which is trained with the generator simultaneously. The saliency network has two functions: (1) providing constraints for content loss to increase punishment for salient regions, and (2) supplying saliency features to generator to produce coherent results. Moreover, two novel losses are proposed to optimize the generator and saliency networks. The proposed method preserves the details on important salient regions and improves the total image perceptual quality. Qualitative and quantitative comparisons against several leading prior methods demonstrates the superiority of our method.
本文提出了一种有效处理显著区域的解决方案,用于在未配对数据集之间进行风格迁移。最近,生成对抗网络(GAN)已经展示了在没有配对示例的情况下将图像从源域X转换到目标域Y的潜力。然而,这种转换不能保证生成高感知质量的结果。现有的风格迁移方法在相对均匀的内容上效果良好,但它们往往无法捕捉总是属于显著区域的几何或结构模式。即使每个单独的区域都被正确地转换为目标风格,结构化区域中的细节损失和平滑区域中不期望的伪影也是不可避免的。在本文中,我们提出了SDP-GAN,一种基于GAN的网络,用于解决此类问题,同时生成令人满意的风格迁移结果。我们引入了一个显著性网络,它与生成器同时进行训练。显著性网络有两个功能:(1)为内容损失提供约束,以增加对显著区域的惩罚;(2)向生成器提供显著性特征,以产生连贯的结果。此外,还提出了两种新颖的损失来优化生成器和显著性网络。所提出的方法保留了重要显著区域的细节,并提高了整体图像的感知质量。与几种领先的现有方法进行的定性和定量比较证明了我们方法的优越性。