• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于图像美学评估的带有污染标签的深度主动学习

Deep Active Learning with Contaminated Tags for Image Aesthetics Assessment.

作者信息

Liu Zhenguang, Wang Zepeng, Yao Yiyang, Zhang Luming, Shao Ling

出版信息

IEEE Trans Image Process. 2018 Apr 18. doi: 10.1109/TIP.2018.2828326.

DOI:10.1109/TIP.2018.2828326
PMID:29993633
Abstract

Image aesthetic quality assessment has becoming an indispensable technique that facilitates a variety of image applications, e.g., photo retargeting and non-realistic rendering. Conventional approaches suffer from the following limitations: 1) the inefficiency of semantically describing images due to the inherent tag noise and incompletion, 2) the difficulty of accurately reflecting how humans actively perceive various regions inside each image, and 3) the challenge of incorporating the aesthetic experiences of multiple users. To solve these problems, we propose a novel semi-supervised deep active learning (SDAL) algorithm, which discovers how humans perceive semantically important regions from a large quantity of images partially assigned with contaminated tags. More specifically, as humans usually attend to the foreground objects before understanding them, we extract a succinct set of BING (binarized normed gradients) [60]-based object patches from each image. To simulate human visual perception, we propose SDAL which hierarchically learns human gaze shifting path (GSP) by sequentially linking semantically important object patches from each scenery. Noticeably, SDLA unifies the semantically important regions discovery and deep GSP feature learning into a principled framework, wherein only a small proportion of tagged images are adopted. Moreover, based on the sparsity penalty, SDLA can optimally abandon the noisy or redundant low-level image features. Finally, by leveraging the deeply-learned GSP features, a probabilistic model is developed for image aesthetics assessment, where the experience of multiple professional photographers can be encoded. Besides, auxiliary quality-related features can be conveniently integrated into our probabilistic model. Comprehensive experiments on a series of benchmark image sets have demonstrated the superiority of our method. As a byproduct, eye tracking experiments have shown that GSPs generated by our SDAL are about 93% consistent with real human gaze shifting paths.

摘要

图像美学质量评估已成为一种不可或缺的技术,它推动了各种图像应用,例如照片重定目标尺寸和非真实感渲染。传统方法存在以下局限性:1)由于固有的标签噪声和不完整性,在语义上描述图像效率低下;2)难以准确反映人类如何主动感知每张图像内的各个区域;3)难以纳入多个用户的美学体验。为了解决这些问题,我们提出了一种新颖的半监督深度主动学习(SDAL)算法,该算法从大量带有受污染标签的部分图像中发现人类如何感知语义上重要的区域。更具体地说,由于人类通常在理解前景对象之前会先关注它们,我们从每张图像中提取了一组基于BING(二值化规范梯度)[60]的简洁对象块。为了模拟人类视觉感知,我们提出了SDAL,它通过依次链接每个场景中语义上重要的对象块,分层学习人类注视转移路径(GSP)。值得注意的是,SDLA将语义上重要区域的发现和深度GSP特征学习统一到一个有原则的框架中,其中仅采用一小部分带标签的图像。此外,基于稀疏性惩罚,SDLA可以最佳地舍弃噪声或冗余的低级图像特征。最后,通过利用深度学习的GSP特征,开发了一种用于图像美学评估的概率模型,其中可以编码多个专业摄影师的经验。此外,与质量相关的辅助特征可以方便地集成到我们的概率模型中。在一系列基准图像集上进行的综合实验证明了我们方法的优越性。作为一个副产品,眼动追踪实验表明,我们的SDAL生成的GSP与真实人类注视转移路径的一致性约为93%。

相似文献

1
Deep Active Learning with Contaminated Tags for Image Aesthetics Assessment.用于图像美学评估的带有污染标签的深度主动学习
IEEE Trans Image Process. 2018 Apr 18. doi: 10.1109/TIP.2018.2828326.
2
Bioinspired Scene Classification by Deep Active Learning With Remote Sensing Applications.基于深度学习的主动学习在遥感场景分类中的应用
IEEE Trans Cybern. 2022 Jul;52(7):5682-5694. doi: 10.1109/TCYB.2020.2981480. Epub 2022 Jul 4.
3
Scene Categorization Using Deeply Learned Gaze Shifting Kernel.基于深度学习的注视转移核的场景分类。
IEEE Trans Cybern. 2019 Jun;49(6):2156-2167. doi: 10.1109/TCYB.2018.2820731. Epub 2018 May 11.
4
Massive-Scale Aerial Photo Categorization by Cross-Resolution Visual Perception Enhancement.大规模航空影像分类的跨分辨率视觉感知增强方法
IEEE Trans Neural Netw Learn Syst. 2022 Aug;33(8):4017-4030. doi: 10.1109/TNNLS.2021.3055548. Epub 2022 Aug 3.
5
Perceptually Aware Image Retargeting for Mobile Devices.移动端有感知的图像重定向。
IEEE Trans Image Process. 2018 May;27(5):2301-2313. doi: 10.1109/TIP.2017.2779272.
6
Engineering Deep Representations for Modeling Aesthetic Perception.工程化深度表示以建模审美感知。
IEEE Trans Cybern. 2018 Nov;48(11):3092-3104. doi: 10.1109/TCYB.2017.2758350. Epub 2017 Dec 11.
7
Community-Aware Photo Quality Evaluation by Deeply Encoding Human Perception.基于深度学习的人类感知编码的社区感知图像质量评价
IEEE Trans Cybern. 2022 May;52(5):3136-3146. doi: 10.1109/TCYB.2019.2937319. Epub 2022 May 19.
8
Scene Categorization by Deeply Learning Gaze Behavior in a Semisupervised Context.在半监督环境下通过深度学习注视行为进行场景分类
IEEE Trans Cybern. 2021 Aug;51(8):4265-4276. doi: 10.1109/TCYB.2019.2913016. Epub 2021 Aug 4.
9
Aesthetics-Guided Graph Clustering With Absent Modalities Imputation.
IEEE Trans Image Process. 2019 Jul;28(7):3462-3476. doi: 10.1109/TIP.2019.2897940. Epub 2019 Feb 6.
10
Actively learning human gaze shifting paths for semantics-aware photo cropping.主动学习人类注视转移路径以实现语义感知的照片裁剪。
IEEE Trans Image Process. 2014 May;23(5):2235-45. doi: 10.1109/TIP.2014.2311658.

引用本文的文献

1
And the nominees are: Using design-awards datasets to build computational aesthetic evaluation model.提名的有:利用设计奖项数据集构建计算美学评估模型。
PLoS One. 2020 Jan 21;15(1):e0227754. doi: 10.1371/journal.pone.0227754. eCollection 2020.