• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

运动力:解码第一人称视频中活动的控制力。

Force from Motion: Decoding Control Force of Activity in a First-Person Video.

出版信息

IEEE Trans Pattern Anal Mach Intell. 2020 Mar;42(3):622-635. doi: 10.1109/TPAMI.2018.2883327. Epub 2018 Nov 26.

DOI:10.1109/TPAMI.2018.2883327
PMID:30489262
Abstract

A first-person video delivers what the camera wearer (actor) experiences through physical interactions with surroundings. In this paper, we focus on a problem of Force from Motion-estimating the active force and torque exerted by the actor to drive her/his activity-from a first-person video. We use two physical cues inherited in the first-person video. (1) Ego-motion: the camera motion is generated by a resultant of force interactions, which allows us to understand the effect of the active force using Newtonian mechanics. (2) Visual semantics: the first-person visual scene is deployed to afford the actor's activity, which is indicative of the physical context of the activity. We estimate the active force and torque using a dynamical system that can describe the transition (dynamics) of the actor's physical state (position, orientation, and linear/angular momentum) where the latent physical state is indirectly observed by the first-person video. We approximate the physical state with the 3D camera trajectory that is reconstructed up to scale and orientation. The absolute scale factor and gravitation field are learned from the ego-motion and visual semantics of the first-person video. Inspired by an optimal control theory, we solve the dynamical system by minimizing reprojection error. Our method shows quantitatively equivalent reconstruction comparing to IMU measurements in terms of gravity and scale recovery and outperforms the methods based on 2D optical flow for an active action recognition task. We apply our method to first-person videos of mountain biking, urban bike racing, skiing, speedflying with parachute, and wingsuit flying where inertial measurements are not accessible.

摘要

第一人称视频传达了摄像机佩戴者(演员)通过与周围环境的物理交互所体验到的内容。在本文中,我们专注于从第一人称视频中估计演员施加的力和扭矩的力从运动估计问题。我们使用第一人称视频中继承的两个物理线索。(1) 自身运动:摄像机运动是由力相互作用的合力产生的,这使我们能够使用牛顿力学理解主动力的效果。(2) 视觉语义:第一人称视觉场景用于提供演员的活动,这表明了活动的物理背景。我们使用可以描述演员物理状态(位置、方向和线性/角动量)过渡(动力学)的动力系统来估计主动力和扭矩,其中潜在的物理状态通过第一人称视频间接观察。我们使用重建到比例和方向的 3D 摄像机轨迹来近似物理状态。绝对比例因子和重力场是从第一人称视频的自身运动和视觉语义中学习的。受最优控制理论的启发,我们通过最小化重投影误差来求解动力系统。与基于 IMU 测量的方法相比,我们的方法在重力和比例恢复方面具有定量等效的重建效果,并且在主动动作识别任务中优于基于 2D 光流的方法。我们将我们的方法应用于山地自行车、城市自行车赛车、滑雪、带降落伞的速度飞行和翼装飞行的第一人称视频中,在这些视频中无法访问惯性测量。

相似文献

1
Force from Motion: Decoding Control Force of Activity in a First-Person Video.运动力:解码第一人称视频中活动的控制力。
IEEE Trans Pattern Anal Mach Intell. 2020 Mar;42(3):622-635. doi: 10.1109/TPAMI.2018.2883327. Epub 2018 Nov 26.
2
Desktop Action Recognition From First-Person Point-of-View.基于第一人称视角的桌面行为识别。
IEEE Trans Cybern. 2019 May;49(5):1616-1628. doi: 10.1109/TCYB.2018.2806381. Epub 2018 Feb 27.
3
A general framework for tracking multiple people from a moving camera.从移动摄像机跟踪多个人的通用框架。
IEEE Trans Pattern Anal Mach Intell. 2013 Jul;35(7):1577-91. doi: 10.1109/TPAMI.2012.248.
4
Recognizing physical activity from ego-motion of a camera.从相机的自我运动中识别身体活动。
Annu Int Conf IEEE Eng Med Biol Soc. 2010;2010:5569-72. doi: 10.1109/IEMBS.2010.5626794.
5
Nonlinear Low-Rank Matrix Completion for Human Motion Recovery.基于非线性低秩矩阵填充的人体运动恢复。
IEEE Trans Image Process. 2018 Jun;27(6):3011-3024. doi: 10.1109/TIP.2018.2812100. Epub 2018 Mar 5.
6
Learning and Inferring "Dark Matter" and Predicting Human Intents and Trajectories in Videos.学习和推断“暗物质”,并预测视频中的人类意图和轨迹。
IEEE Trans Pattern Anal Mach Intell. 2018 Jul;40(7):1639-1652. doi: 10.1109/TPAMI.2017.2728788. Epub 2017 Jul 19.
7
Explicit modeling of human-object interactions in realistic videos.真实视频中人类-物体交互的显式建模。
IEEE Trans Pattern Anal Mach Intell. 2013 Apr;35(4):835-48. doi: 10.1109/TPAMI.2012.175.
8
A semantic autonomous video surveillance system for dense camera networks in Smart Cities.一种用于智慧城市中密集型摄像机网络的语义自主视频监控系统。
Sensors (Basel). 2012;12(8):10407-29. doi: 10.3390/s120810407. Epub 2012 Aug 2.
9
Scalable Semi-Automatic Annotation for Multi-Camera Person Tracking.多摄像机人像跟踪的可扩展半自动标注
IEEE Trans Image Process. 2016 May;25(5):2259-74. doi: 10.1109/TIP.2016.2542021.
10
Ego-Surfing: Person Localization in First-Person Videos Using Ego-Motion Signatures.自我冲浪:利用自我运动特征进行第一人称视频中的人物定位。
IEEE Trans Pattern Anal Mach Intell. 2018 Nov;40(11):2749-2761. doi: 10.1109/TPAMI.2017.2771767. Epub 2017 Nov 9.