Li Haoxuan, Sheng Bin, Li Ping, Ali Riaz, Chen C L Philip
IEEE Trans Image Process. 2021;30:8526-8539. doi: 10.1109/TIP.2021.3117061. Epub 2021 Oct 13.
Given a target grayscale image and a reference color image, exemplar-based image colorization aims to generate a visually natural-looking color image by transforming meaningful color information from the reference image to the target image. It remains a challenging problem due to the differences in semantic content between the target image and the reference image. In this paper, we present a novel globally and locally semantic colorization method called exemplar-based conditional broad-GAN, a broad generative adversarial network (GAN) framework, to deal with this limitation. Our colorization framework is composed of two sub-networks: the match sub-net and the colorization sub-net. We reconstruct the target image with a dictionary-based sparse representation in the match sub-net, where the dictionary consists of features extracted from the reference image. To enforce global-semantic and local-structure self-similarity constraints, global-local affinity energy is explored to constrain the sparse representation for matching consistency. Then, the matching information of the match sub-net is fed into the colorization sub-net as the perceptual information of the conditional broad-GAN to facilitate the personalized results. Finally, inspired by the observation that a broad learning system is able to extract semantic features efficiently, we further introduce a broad learning system into the conditional GAN and propose a novel loss, which substantially improves the training stability and the semantic similarity between the target image and the ground truth. Extensive experiments have shown that our colorization approach outperforms the state-of-the-art methods, both perceptually and semantically.
给定一个目标灰度图像和一个参考彩色图像,基于样本的图像上色旨在通过将参考图像中有意义的颜色信息转换到目标图像,来生成视觉上自然的彩色图像。由于目标图像和参考图像在语义内容上存在差异,这仍然是一个具有挑战性的问题。在本文中,我们提出了一种新颖的全局和局部语义上色方法,称为基于样本的条件广义生成对抗网络(exemplar-based conditional broad-GAN),这是一种广义生成对抗网络(GAN)框架,以应对这一局限性。我们的上色框架由两个子网络组成:匹配子网络和上色子网络。我们在匹配子网络中使用基于字典的稀疏表示来重建目标图像,其中字典由从参考图像中提取的特征组成。为了强制全局语义和局部结构自相似性约束,探索全局-局部亲和能量来约束稀疏表示以实现匹配一致性。然后,将匹配子网络的匹配信息作为条件广义生成对抗网络的感知信息输入到上色子网络中,以促进个性化结果。最后,受广义学习系统能够有效提取语义特征这一观察结果的启发,我们进一步将广义学习系统引入到条件生成对抗网络中,并提出了一种新颖的损失函数,这极大地提高了训练稳定性以及目标图像与真实图像之间的语义相似度。大量实验表明,我们的上色方法在感知和语义方面均优于现有方法。