• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

以最少标注工作量实现静止图像中的动作识别

Action Recognition in Still Images With Minimum Annotation Efforts.

作者信息

Do Minh N

出版信息

IEEE Trans Image Process. 2016 Nov;25(11):5479-5490. doi: 10.1109/TIP.2016.2605305. Epub 2016 Sep 1.

DOI:10.1109/TIP.2016.2605305
PMID:27608461
Abstract

We focus on the problem of still image-based human action recognition, which essentially involves making prediction by analyzing human poses and their interaction with objects in the scene. Besides image-level action labels (e.g., riding, phoning), during both training and testing stages, existing works usually require additional input of human bounding boxes to facilitate the characterization of the underlying human-object interactions. We argue that this additional input requirement might severely discourage potential applications and is not very necessary. To this end, a systematic approach was developed in this paper to address this challenging problem of minimum annotation efforts, i.e., to perform recognition in the presence of only image-level action labels in the training stage. Experimental results on three benchmark data sets demonstrate that compared with the state-of-the-art methods that have privileged access to additional human bounding-box annotations, our approach achieves comparable or even superior recognition accuracy using only action annotations in training. Interestingly, as a by-product in many cases, our approach is able to segment out the precise regions of underlying human-object interactions.

摘要

我们专注于基于静止图像的人类动作识别问题,该问题本质上涉及通过分析人体姿势及其与场景中物体的交互来进行预测。除了图像级别的动作标签(例如,骑行、打电话)之外,在训练和测试阶段,现有工作通常需要额外输入人体边界框,以促进对潜在的人体 - 物体交互的表征。我们认为,这种额外的输入要求可能会严重阻碍潜在应用,并且并非十分必要。为此,本文开发了一种系统方法来解决这个具有挑战性的最小标注工作量问题,即在训练阶段仅存在图像级动作标签的情况下进行识别。在三个基准数据集上的实验结果表明,与那些能够获取额外人体边界框标注的最先进方法相比,我们的方法在训练中仅使用动作标注就能达到相当甚至更高的识别准确率。有趣的是,在许多情况下,作为副产品,我们的方法能够分割出潜在人体 - 物体交互的精确区域。

相似文献

1
Action Recognition in Still Images With Minimum Annotation Efforts.以最少标注工作量实现静止图像中的动作识别
IEEE Trans Image Process. 2016 Nov;25(11):5479-5490. doi: 10.1109/TIP.2016.2605305. Epub 2016 Sep 1.
2
AutoBD: Automated Bi-Level Description for Scalable Fine-Grained Visual Categorization.AutoBD:用于可扩展细粒度视觉分类的自动双层次描述。
IEEE Trans Image Process. 2018;27(1):10-23. doi: 10.1109/TIP.2017.2751960.
3
Recognizing Actions Through Action-Specific Person Detection.通过特定于动作的人物检测来识别动作。
IEEE Trans Image Process. 2015 Nov;24(11):4422-32. doi: 10.1109/TIP.2015.2465147. Epub 2015 Aug 5.
4
Weakly Supervised Fine-Grained Categorization With Part-Based Image Representation.基于部件的图像表示的弱监督细粒度分类。
IEEE Trans Image Process. 2016 Apr;25(4):1713-25. doi: 10.1109/TIP.2016.2531289. Epub 2016 Feb 18.
5
Weakly Supervised Large Scale Object Localization with Multiple Instance Learning and Bag Splitting.基于多示例学习和 Bag Splitting 的弱监督大规模目标定位。
IEEE Trans Pattern Anal Mach Intell. 2016 Feb;38(2):405-16. doi: 10.1109/TPAMI.2015.2456908.
6
Semantic pyramids for gender and action recognition.用于性别和动作识别的语义金字塔。
IEEE Trans Image Process. 2014 Aug;23(8):3633-45. doi: 10.1109/TIP.2014.2331759. Epub 2014 Jun 18.
7
Training Robust Object Detectors From Noisy Category Labels and Imprecise Bounding Boxes.从噪声类别标签和不精确边界框中训练鲁棒目标检测器。
IEEE Trans Image Process. 2021;30:5782-5792. doi: 10.1109/TIP.2021.3085208. Epub 2021 Jun 23.
8
Simultaneously Discovering and Localizing Common Objects in Wild Images.在野外图像中同时发现和定位常见对象。
IEEE Trans Image Process. 2018 Sep;27(9):4503-4515. doi: 10.1109/TIP.2018.2839901.
9
Saliency-Aware Nonparametric Foreground Annotation Based on Weakly Labeled Data.基于弱标注数据的显著度感知非参数前景标注
IEEE Trans Neural Netw Learn Syst. 2016 Jun;27(6):1253-65. doi: 10.1109/TNNLS.2015.2488637. Epub 2015 Oct 26.
10
Selective Convolutional Descriptor Aggregation for Fine-Grained Image Retrieval.选择性卷积描述符聚合用于细粒度图像检索。
IEEE Trans Image Process. 2017 Jun;26(6):2868-2881. doi: 10.1109/TIP.2017.2688133. Epub 2017 Mar 27.

引用本文的文献

1
Recognition of human action for scene understanding using world cup optimization and transfer learning approach.使用世界杯优化和迁移学习方法进行场景理解的人类行为识别
PeerJ Comput Sci. 2023 May 23;9:e1396. doi: 10.7717/peerj-cs.1396. eCollection 2023.