• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于跨域少样本动作识别的成对注意力对抗时空网络 - R2

A Pairwise Attentive Adversarial Spatiotemporal Network for Cross-Domain Few-Shot Action Recognition-R2.

作者信息

Gao Zan, Guo Leming, Guan Weili, Liu An-An, Ren Tongwei, Chen Shengyong

出版信息

IEEE Trans Image Process. 2021;30:767-782. doi: 10.1109/TIP.2020.3038372. Epub 2020 Dec 4.

DOI:10.1109/TIP.2020.3038372
PMID:33232234
Abstract
  • Action recognition is a popular research topic in the computer vision and machine learning domains. Although many action recognition methods have been proposed, only a few researchers have focused on cross-domain few-shot action recognition, which must often be performed in real security surveillance. Since the problems of action recognition, domain adaptation, and few-shot learning need to be simultaneously solved, the cross-domain few-shot action recognition task is a challenging problem. To solve these issues, in this work, we develop a novel end-to-end pairwise attentive adversarial spatiotemporal network (PASTN) to perform the cross-domain few-shot action recognition task, in which spatiotemporal information acquisition, few-shot learning, and video domain adaptation are realised in a unified framework. Specifically, the Resnet-50 network is selected as the backbone of the PASTN, and a 3D convolution block is embedded in the top layer of the 2D CNN (ResNet-50) to capture the spatiotemporal representations. Moreover, a novel attentive adversarial network architecture is designed to align the spatiotemporal dynamics actions with higher domain discrepancies. In addition, the pairwise margin discrimination loss is designed for the pairwise network architecture to improve the discrimination of the learned domain-invariant spatiotemporal feature. The results of extensive experiments performed on three public benchmarks of the cross-domain action recognition datasets, including SDAI Action I, SDAI Action II and UCF50-OlympicSport, demonstrate that the proposed PASTN can significantly outperform the state-of-the-art cross-domain action recognition methods in terms of both the accuracy and computational time. Even when only two labelled training samples per category are considered in the office1 scenario of the SDAI Action I dataset, the accuracy of the PASTN is improved by 6.1%, 10.9%, 16.8%, and 14% compared to that of the TAN , TemporalPooling, I3D, and P3D methods, respectively.
摘要

动作识别是计算机视觉和机器学习领域中一个热门的研究课题。尽管已经提出了许多动作识别方法,但只有少数研究人员关注跨域少样本动作识别,而这一任务在实际的安全监控中经常需要执行。由于动作识别、域适应和少样本学习的问题需要同时解决,跨域少样本动作识别任务是一个具有挑战性的问题。为了解决这些问题,在这项工作中,我们开发了一种新颖的端到端成对注意力对抗时空网络(PASTN)来执行跨域少样本动作识别任务,其中时空信息获取、少样本学习和视频域适应在一个统一的框架中实现。具体来说,选择Resnet-50网络作为PASTN的主干,并在2D CNN(ResNet-50)的顶层嵌入一个3D卷积块来捕捉时空表示。此外,设计了一种新颖的注意力对抗网络架构,以使具有更高域差异的时空动态动作对齐。另外,为成对网络架构设计了成对边际判别损失,以提高所学域不变时空特征的判别能力。在跨域动作识别数据集的三个公共基准上进行的大量实验结果,包括SDAI Action I、SDAI Action II和UCF50-OlympicSport,表明所提出的PASTN在准确性和计算时间方面都能显著优于当前的跨域动作识别方法。即使在SDAI Action I数据集的office1场景中,每类仅考虑两个标记的训练样本时,与TAN、TemporalPooling、I3D和P3D方法相比,PASTN的准确率分别提高了6.1%、10.9%、16.8%和14%。

相似文献

1
A Pairwise Attentive Adversarial Spatiotemporal Network for Cross-Domain Few-Shot Action Recognition-R2.用于跨域少样本动作识别的成对注意力对抗时空网络 - R2
IEEE Trans Image Process. 2021;30:767-782. doi: 10.1109/TIP.2020.3038372. Epub 2020 Dec 4.
2
A Novel Multiple-View Adversarial Learning Network for Unsupervised Domain Adaptation Action Recognition.一种新颖的基于多视角对抗学习的无监督域自适应动作识别网络。
IEEE Trans Cybern. 2022 Dec;52(12):13197-13211. doi: 10.1109/TCYB.2021.3105637. Epub 2022 Nov 18.
3
Pairwise Two-Stream ConvNets for Cross-Domain Action Recognition With Small Data.用于小数据跨域动作识别的成对双流卷积神经网络
IEEE Trans Neural Netw Learn Syst. 2022 Mar;33(3):1147-1161. doi: 10.1109/TNNLS.2020.3041018. Epub 2022 Feb 28.
4
Cross-Modal Contrastive Learning Network for Few-Shot Action Recognition.用于少样本动作识别的跨模态对比学习网络
IEEE Trans Image Process. 2024;33:1257-1271. doi: 10.1109/TIP.2024.3354104. Epub 2024 Feb 13.
5
Self-Paced Collaborative and Adversarial Network for Unsupervised Domain Adaptation.自定步幅协作与对抗网络的无监督域自适应。
IEEE Trans Pattern Anal Mach Intell. 2021 Jun;43(6):2047-2061. doi: 10.1109/TPAMI.2019.2962476. Epub 2021 May 11.
6
Generalizable and Discriminative Representations for Adversarially Robust Few-Shot Learning.用于对抗鲁棒少样本学习的可泛化和判别性表示
IEEE Trans Neural Netw Learn Syst. 2025 Mar;36(3):5480-5493. doi: 10.1109/TNNLS.2024.3379172. Epub 2025 Feb 28.
7
Modality independent adversarial network for generalized zero shot image classification.模态无关对抗网络的广义零样本图像分类。
Neural Netw. 2021 Feb;134:11-22. doi: 10.1016/j.neunet.2020.11.007. Epub 2020 Nov 21.
8
Spatiotemporal Interaction Residual Networks with Pseudo3D for Video Action Recognition.基于伪 3D 的时空交互残差网络的视频动作识别。
Sensors (Basel). 2020 Jun 1;20(11):3126. doi: 10.3390/s20113126.
9
Cross-Domain Facial Expression Recognition: A Unified Evaluation Benchmark and Adversarial Graph Learning.跨领域面部表情识别:统一的评估基准与对抗图学习。
IEEE Trans Pattern Anal Mach Intell. 2022 Dec;44(12):9887-9903. doi: 10.1109/TPAMI.2021.3131222. Epub 2022 Nov 7.
10
Multi-label zero-shot human action recognition via joint latent ranking embedding.基于联合潜在排序嵌入的多标签零镜头人体动作识别。
Neural Netw. 2020 Feb;122:1-23. doi: 10.1016/j.neunet.2019.09.029. Epub 2019 Oct 21.

引用本文的文献

1
Parallel multi-stage rectification networks for 3D skeleton-based motion prediction.用于基于3D骨架的运动预测的并行多阶段整流网络。
Sci Rep. 2024 Oct 30;14(1):26058. doi: 10.1038/s41598-024-75782-7.
2
Video action recognition collaborative learning with dynamics via PSO-ConvNet Transformer.基于粒子群优化卷积神经网络变压器的带动力学的视频动作识别协同学习
Sci Rep. 2023 Sep 5;13(1):14624. doi: 10.1038/s41598-023-39744-9.
3
An initial prediction and fine-tuning model based on improving GCN for 3D human motion prediction.一种基于改进图卷积网络(GCN)的用于3D人体运动预测的初始预测和微调模型。
Front Comput Neurosci. 2023 Apr 5;17:1145209. doi: 10.3389/fncom.2023.1145209. eCollection 2023.