• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于联合图分解和节点标注的人际交互理解

Human Interaction Understanding With Joint Graph Decomposition and Node Labeling.

作者信息

Wang Zhenhua, Ge Jinchao, Guo Dongyan, Zhang Jianhua, Lei Yanjing, Chen Shengyong

出版信息

IEEE Trans Image Process. 2021;30:6240-6254. doi: 10.1109/TIP.2021.3093383. Epub 2021 Jul 12.

DOI:10.1109/TIP.2021.3093383
PMID:34224352
Abstract

The task of human interaction understanding involves both recognizing the action of each individual in the scene and decoding the interaction relationship among people, which is useful to a series of vision applications such as camera surveillance, video-based sports analysis and event retrieval. This paper divides the task into two problems including grouping people into clusters and assigning labels to each of them, and presents an approach to solving these problems in a joint manner. Our method does not assume the number of groups is known beforehand as this will substantially restrict its application. With the observation that the two challenges are highly correlated, the key idea is to model the pairwise interacting relations among people via a complete graph and its associated energy function such that the labeling and grouping problems are translated into the minimization of the energy function. We implement this joint framework by fusing both deep features and rich contextual cues, and learn the fusion parameters from data. An alternating search algorithm is developed in order to efficiently solve the associated inference problem. By combining the grouping and labeling results obtained with our method, we are able to achieve the semantic-level understanding of human interactions. Extensive experiments are performed to qualitatively and quantitatively evaluate the effectiveness of our approach, which outperforms state-of-the-art methods on several important benchmarks. An ablation study is also performed to verify the effectiveness of different modules within our approach.

摘要

理解人类交互行为的任务包括识别场景中每个人的动作以及解读人与人之间的交互关系,这对于一系列视觉应用都很有用,如摄像头监控、基于视频的体育分析和事件检索。本文将该任务分为两个问题,即把人分组为不同的簇并为每个簇分配标签,并提出一种以联合方式解决这些问题的方法。我们的方法不假定预先知道组的数量,因为这会严重限制其应用。基于这两个挑战高度相关的观察,关键思想是通过一个完全图及其相关的能量函数对人与人之间的成对交互关系进行建模,从而将标签和分组问题转化为能量函数的最小化问题。我们通过融合深度特征和丰富的上下文线索来实现这个联合框架,并从数据中学习融合参数。为了有效解决相关的推理问题,开发了一种交替搜索算法。通过结合使用我们的方法获得的分组和标签结果,我们能够实现对人类交互行为的语义级理解。进行了大量实验,从定性和定量两方面评估我们方法的有效性,在几个重要基准测试中,该方法优于现有方法。还进行了消融研究,以验证我们方法中不同模块的有效性。

相似文献

1
Human Interaction Understanding With Joint Graph Decomposition and Node Labeling.基于联合图分解和节点标注的人际交互理解
IEEE Trans Image Process. 2021;30:6240-6254. doi: 10.1109/TIP.2021.3093383. Epub 2021 Jul 12.
2
Graph Embedded Extreme Learning Machine.图嵌入极限学习机。
IEEE Trans Cybern. 2016 Jan;46(1):311-24. doi: 10.1109/TCYB.2015.2401973. Epub 2015 Mar 2.
3
Close Human Interaction Recognition Using Patch-Aware Models.基于补丁感知模型的近距人类交互识别
IEEE Trans Image Process. 2016 Jan;25(1):167-78. doi: 10.1109/TIP.2015.2498410. Epub 2015 Nov 5.
4
Animated pose templates for modeling and detecting human actions.用于建模和检测人体动作的动画姿势模板。
IEEE Trans Pattern Anal Mach Intell. 2014 Mar;36(3):436-52. doi: 10.1109/TPAMI.2013.144.
5
Learning a Deep Model for Human Action Recognition from Novel Viewpoints.从新视角学习人类动作识别的深度模型。
IEEE Trans Pattern Anal Mach Intell. 2018 Mar;40(3):667-681. doi: 10.1109/TPAMI.2017.2691768. Epub 2017 Apr 6.
6
Force from Motion: Decoding Control Force of Activity in a First-Person Video.运动力:解码第一人称视频中活动的控制力。
IEEE Trans Pattern Anal Mach Intell. 2020 Mar;42(3):622-635. doi: 10.1109/TPAMI.2018.2883327. Epub 2018 Nov 26.
7
Multitask Non-Autoregressive Model for Human Motion Prediction.多任务非自回归人体运动预测模型。
IEEE Trans Image Process. 2021;30:2562-2574. doi: 10.1109/TIP.2020.3038362. Epub 2021 Feb 5.
8
NTU RGB+D 120: A Large-Scale Benchmark for 3D Human Activity Understanding.NTU RGB+D 120:用于三维人体活动理解的大规模基准测试。
IEEE Trans Pattern Anal Mach Intell. 2020 Oct;42(10):2684-2701. doi: 10.1109/TPAMI.2019.2916873. Epub 2019 May 14.
9
Multi-view human activity recognition in distributed camera sensor networks.分布式摄像机传感器网络中的多视角人体活动识别。
Sensors (Basel). 2013 Jul 8;13(7):8750-70. doi: 10.3390/s130708750.
10
Discovering motion primitives for unsupervised grouping and one-shot learning of human actions, gestures, and expressions.发现运动基元,用于人类动作、手势和表情的无监督分组和一次性学习。
IEEE Trans Pattern Anal Mach Intell. 2013 Jul;35(7):1635-48. doi: 10.1109/TPAMI.2012.253.

引用本文的文献

1
A Point-Cloud Segmentation Network Based on SqueezeNet and Time Series for Plants.一种基于SqueezeNet和时间序列的植物点云分割网络。
J Imaging. 2023 Nov 23;9(12):258. doi: 10.3390/jimaging9120258.