Zhou Zheng, Wu Yue, Zhou Yicong
IEEE Trans Neural Netw Learn Syst. 2024 Nov;35(11):16845-16856. doi: 10.1109/TNNLS.2023.3298383. Epub 2024 Oct 29.
Arbitrary style transfer (AST) has garnered considerable attention for its ability to transfer styles infinitely. Although existing methods have achieved impressive results, they may overlook style consistencies and fail to capture crucial style patterns, leading to inconsistent style transfer (ST) caused by minor disturbances. To tackle this issue, we conduct a mathematical analysis of inconsistent ST and develop a style inconsistency measure (SIM) to quantify the inconsistencies between generated images. Moreover, we propose a consistent AST (CAST) framework that effectively captures and transfers essential style features into content images. The proposed CAST framework incorporates an intersection-of-union-preserving crop (IoUPC) module to obtain style pairs with minor disturbance, a self-attention (SA) module to learn the crucial style features, and a style inconsistency loss regularization (SILR) to facilitate consistent feature learning for consistent stylization. Our proposed framework not only provides an optimal solution for consistent ST but also outperforms existing methods when embedded into the CAST framework. Extensive experiments demonstrate that the proposed CAST framework can effectively transfer style patterns while preserving consistency and achieve the state-of-the-art performance.
任意风格迁移(AST)因其能够无限地迁移风格而备受关注。尽管现有方法已经取得了令人瞩目的成果,但它们可能会忽略风格的一致性,无法捕捉关键的风格模式,从而导致由微小干扰引起的不一致风格迁移(ST)。为了解决这个问题,我们对不一致的ST进行了数学分析,并开发了一种风格不一致度量(SIM)来量化生成图像之间的不一致性。此外,我们提出了一种一致的AST(CAST)框架,该框架有效地捕捉基本风格特征并将其迁移到内容图像中。所提出的CAST框架包含一个保留交并比的裁剪(IoUPC)模块,以获得干扰较小的风格对;一个自注意力(SA)模块,以学习关键的风格特征;以及一个风格不一致损失正则化(SILR),以促进一致风格化的一致特征学习。我们提出的框架不仅为一致的ST提供了一个最优解,而且在嵌入到CAST框架中时优于现有方法。大量实验表明,所提出的CAST框架能够有效地迁移风格模式,同时保持一致性,并实现了当前的最优性能。