• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于图像生成的具有语义感知上采样的局部和全局生成对抗网络

Local and Global GANs With Semantic-Aware Upsampling for Image Generation.

作者信息

Tang Hao, Shao Ling, Torr Philip H S, Sebe Nicu

出版信息

IEEE Trans Pattern Anal Mach Intell. 2023 Jan;45(1):768-784. doi: 10.1109/TPAMI.2022.3155989. Epub 2022 Dec 5.

DOI:10.1109/TPAMI.2022.3155989
PMID:35263249
Abstract

In this paper, we address the task of semantic-guided image generation. One challenge common to most existing image-level generation methods is the difficulty in generating small objects and detailed local textures. To address this, in this work we consider generating images using local context. As such, we design a local class-specific generative network using semantic maps as guidance, which separately constructs and learns subgenerators for different classes, enabling it to capture finer details. To learn more discriminative class-specific feature representations for the local generation, we also propose a novel classification module. To combine the advantages of both global image-level and local class-specific generation, a joint generation network is designed with an attention fusion module and a dual-discriminator structure embedded. Lastly, we propose a novel semantic-aware upsampling method, which has a larger receptive field and can take far-away pixels that are semantically related for feature upsampling, enabling it to better preserve semantic consistency for instances with the same semantic labels. Extensive experiments on two image generation tasks show the superior performance of the proposed method. State-of-the-art results are established by large margins on both tasks and on nine challenging public benchmarks. The source code and trained models are available at https://github.com/Ha0Tang/LGGAN.

摘要

在本文中,我们探讨语义引导图像生成任务。大多数现有图像级生成方法面临的一个共同挑战是生成小物体和详细局部纹理的困难。为了解决这个问题,在这项工作中我们考虑使用局部上下文来生成图像。因此,我们设计了一个以语义图为指导的局部特定类生成网络,该网络为不同类分别构建并学习子生成器,使其能够捕捉更精细的细节。为了学习用于局部生成的更具判别力的特定类特征表示,我们还提出了一种新颖的分类模块。为了结合全局图像级和局部特定类生成的优点,设计了一个联合生成网络,其中嵌入了注意力融合模块和双判别器结构。最后,我们提出了一种新颖的语义感知上采样方法,它具有更大的感受野,并且可以采用语义相关的远处像素进行特征上采样,从而能够更好地保持具有相同语义标签的实例的语义一致性。在两个图像生成任务上的大量实验表明了所提出方法的卓越性能。在这两个任务以及九个具有挑战性的公共基准上,该方法均大幅领先于当前最优结果。源代码和训练模型可在https://github.com/Ha0Tang/LGGAN获取。

相似文献

1
Local and Global GANs With Semantic-Aware Upsampling for Image Generation.用于图像生成的具有语义感知上采样的局部和全局生成对抗网络
IEEE Trans Pattern Anal Mach Intell. 2023 Jan;45(1):768-784. doi: 10.1109/TPAMI.2022.3155989. Epub 2022 Dec 5.
2
Multi-Channel Attention Selection GANs for Guided Image-to-Image Translation.用于引导图像到图像翻译的多通道注意力选择生成对抗网络
IEEE Trans Pattern Anal Mach Intell. 2023 May;45(5):6055-6071. doi: 10.1109/TPAMI.2022.3212915. Epub 2023 Apr 3.
3
Layout-to-Image Translation With Double Pooling Generative Adversarial Networks.基于双池化生成对抗网络的布局到图像翻译
IEEE Trans Image Process. 2021;30:7903-7913. doi: 10.1109/TIP.2021.3109531. Epub 2021 Sep 20.
4
Unified Generative Adversarial Networks for Controllable Image-to-Image Translation.用于可控图像到图像转换的统一生成对抗网络。
IEEE Trans Image Process. 2020 Sep 11;PP. doi: 10.1109/TIP.2020.3021789.
5
Attention Guided Global Enhancement and Local Refinement Network for Semantic Segmentation.用于语义分割的注意力引导全局增强与局部细化网络
IEEE Trans Image Process. 2022;31:3211-3223. doi: 10.1109/TIP.2022.3166673. Epub 2022 Apr 22.
6
GNDAN: Graph Navigated Dual Attention Network for Zero-Shot Learning.GNDAN:用于零样本学习的图导航双注意力网络。
IEEE Trans Neural Netw Learn Syst. 2024 Apr;35(4):4516-4529. doi: 10.1109/TNNLS.2022.3155602. Epub 2024 Apr 4.
7
Towards a Flexible Semantic Guided Model for Single Image Enhancement and Restoration.迈向用于单图像增强与恢复的灵活语义引导模型。
IEEE Trans Pattern Anal Mach Intell. 2024 Dec;46(12):9921-9939. doi: 10.1109/TPAMI.2024.3432308. Epub 2024 Nov 6.
8
AttentionGAN: Unpaired Image-to-Image Translation Using Attention-Guided Generative Adversarial Networks.AttentionGAN:基于注意力引导生成对抗网络的非配对图像到图像翻译。
IEEE Trans Neural Netw Learn Syst. 2023 Apr;34(4):1972-1987. doi: 10.1109/TNNLS.2021.3105725. Epub 2023 Apr 4.
9
Class-Aware Adversarial Transformers for Medical Image Segmentation.用于医学图像分割的类别感知对抗变压器
Adv Neural Inf Process Syst. 2022 Dec;35:29582-29596.
10
Repurposing GANs for One-Shot Semantic Part Segmentation.GAN 再利用用于一次性语义部分分割。
IEEE Trans Pattern Anal Mach Intell. 2023 Apr;45(4):5114-5125. doi: 10.1109/TPAMI.2022.3201285. Epub 2023 Mar 7.

引用本文的文献

1
UNet-like network fused swin transformer and CNN for semantic image synthesis.用于语义图像合成的类似UNet的网络,融合了Swin Transformer和卷积神经网络。
Sci Rep. 2024 Jul 21;14(1):16761. doi: 10.1038/s41598-024-65585-1.
2
A Review of Performance Prediction Based on Machine Learning in Materials Science.基于机器学习的材料科学性能预测综述
Nanomaterials (Basel). 2022 Aug 26;12(17):2957. doi: 10.3390/nano12172957.