• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种通用风格迁移的统一框架:风格与内容分离。

A Unified Framework for Generalizable Style Transfer: Style and Content Separation.

作者信息

Zhang Yexun, Zhang Ya, Cai Wenbin

出版信息

IEEE Trans Image Process. 2020 Jan 31. doi: 10.1109/TIP.2020.2969081.

DOI:10.1109/TIP.2020.2969081
PMID:32012013
Abstract

Image style transfer has drawn broad attention recently. However, most existing methods aim to explicitly model the transformation between different styles, and the learned model is often not generalizable to new styles. Based on the idea of style and content separation, we here propose a unified style transfer framework that consists of style encoder, content encoder, mixer and decoder. The style encoder and the content encoder are used to extract the style and content representations from the corresponding reference images. The two representations are integrated by the mixer and fed to the decoder, which generates images with the target style and content. Assuming the same encoder could be shared among different styles/contents, the style/content encoder explores a generalizable way to represent style/content information, i.e. the encoders are expected to capture the underlying representation for different styles/contents and generalize to new styles/contents. Training simultaneously with a number of styles and contents, the framework enables building one single transfer network for multiple styles and further leads to a key merit of the framework, i.e. its generalizability to new styles and contents. To evaluate the proposed framework, we apply it to both supervised and unsupervised style transfer, using character typeface transfer and neural style transfer as respective examples. For character typeface transfer, to separate the style features and content features, we leverage the conditional dependence of styles and contents given an image. For neural style transfer, we leverage the statistical information of feature maps in certain layers to represent style. Extensive experimental results have demonstrated the effectiveness and robustness of the proposed methods. Furthermore, models learned under the proposed framework are shown to be better generalizable to new styles and contents.

摘要

图像风格迁移最近引起了广泛关注。然而,大多数现有方法旨在显式地对不同风格之间的转换进行建模,而所学习的模型通常不能推广到新的风格。基于风格与内容分离的思想,我们在此提出一个统一的风格迁移框架,它由风格编码器、内容编码器、混合器和解码器组成。风格编码器和内容编码器用于从相应的参考图像中提取风格和内容表示。这两种表示由混合器进行整合,并馈送到解码器,解码器生成具有目标风格和内容的图像。假设相同的编码器可以在不同的风格/内容之间共享,风格/内容编码器探索一种可推广的方式来表示风格/内容信息,即期望编码器捕获不同风格/内容的潜在表示并推广到新的风格/内容。该框架通过与多种风格和内容同时进行训练,能够构建一个针对多种风格的单一迁移网络,进而带来该框架的一个关键优点,即其对新风格和内容的可推广性。为了评估所提出的框架,我们将其应用于有监督和无监督的风格迁移,分别以字符字体迁移和神经风格迁移为例。对于字符字体迁移,为了分离风格特征和内容特征,我们利用给定图像时风格和内容的条件依赖性。对于神经风格迁移,我们利用某些层中特征图的统计信息来表示风格。大量实验结果证明了所提方法的有效性和鲁棒性。此外,在所提出的框架下学习的模型被证明对新风格和内容具有更好的可推广性。

相似文献

1
A Unified Framework for Generalizable Style Transfer: Style and Content Separation.一种通用风格迁移的统一框架:风格与内容分离。
IEEE Trans Image Process. 2020 Jan 31. doi: 10.1109/TIP.2020.2969081.
2
Gated-GAN: Adversarial Gated Networks for Multi-Collection Style Transfer.门控 GAN:用于多集合风格迁移的对抗门控网络。
IEEE Trans Image Process. 2019 Feb;28(2):546-560. doi: 10.1109/TIP.2018.2869695. Epub 2018 Sep 12.
3
Explicit Filterbank Learning for Neural Image Style Transfer and Image Processing.滤波器组学习在神经图像风格迁移和图像处理中的应用。
IEEE Trans Pattern Anal Mach Intell. 2021 Jul;43(7):2373-2387. doi: 10.1109/TPAMI.2020.2964205. Epub 2021 Jun 9.
4
IFFMStyle: High-Quality Image Style Transfer Using Invalid Feature Filter Modules.IFFM风格:使用无效特征过滤模块的高质量图像风格迁移
Sensors (Basel). 2022 Aug 16;22(16):6134. doi: 10.3390/s22166134.
5
GAN-Based Multi-Style Photo Cartoonization.基于生成对抗网络的多风格照片卡通化
IEEE Trans Vis Comput Graph. 2022 Oct;28(10):3376-3390. doi: 10.1109/TVCG.2021.3067201. Epub 2022 Sep 1.
6
StyleTalk++: A Unified Framework for Controlling the Speaking Styles of Talking Heads.StyleTalk++:用于控制说话人头的说话风格的统一框架。
IEEE Trans Pattern Anal Mach Intell. 2024 Jun;46(6):4331-4347. doi: 10.1109/TPAMI.2024.3357808. Epub 2024 May 7.
7
Image style transfer with collection representation space and semantic-guided reconstruction.基于集合表示空间和语义引导重建的图像风格迁移。
Neural Netw. 2020 Sep;129:123-137. doi: 10.1016/j.neunet.2020.05.028. Epub 2020 Jun 2.
8
A model integrating attention mechanism and generative adversarial network for image style transfer.一种融合注意力机制和生成对抗网络的图像风格迁移模型。
PeerJ Comput Sci. 2024 Sep 20;10:e2332. doi: 10.7717/peerj-cs.2332. eCollection 2024.
9
PSA-Net: Deep learning-based physician style-aware segmentation network for postoperative prostate cancer clinical target volumes.PSA-Net:基于深度学习的医师风格感知分割网络,用于术后前列腺癌临床靶区。
Artif Intell Med. 2021 Nov;121:102195. doi: 10.1016/j.artmed.2021.102195. Epub 2021 Oct 18.
10
Uncorrelated feature encoding for faster image style transfer.非相关特征编码,实现更快的图像风格迁移。
Neural Netw. 2021 Aug;140:148-157. doi: 10.1016/j.neunet.2021.03.007. Epub 2021 Mar 13.