• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

CAM-Vtrans:利用多模态机器人数据的实时运动训练

CAM-Vtrans: real-time sports training utilizing multi-modal robot data.

作者信息

LinLin Hong, Sangheang Lee, GuanTing Song

机构信息

College of Physical Education, Jeonju University, Jeonju, Jeollabuk-do, Republic of Korea.

Gongqing Institute of Science and Technology, Jiujiang, Jiangxi Province, China.

出版信息

Front Neurorobot. 2024 Oct 11;18:1453571. doi: 10.3389/fnbot.2024.1453571. eCollection 2024.

DOI:10.3389/fnbot.2024.1453571
PMID:39463860
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11502466/
Abstract

INTRODUCTION

Assistive robots and human-robot interaction have become integral parts of sports training. However, existing methods often fail to provide real-time and accurate feedback, and they often lack integration of comprehensive multi-modal data.

METHODS

To address these issues, we propose a groundbreaking and innovative approach: CAM-Vtrans-Cross-Attention Multi-modal Visual Transformer. By leveraging the strengths of state-of-the-art techniques such as Visual Transformers (ViT) and models like CLIP, along with cross-attention mechanisms, CAM-Vtrans harnesses the power of visual and textual information to provide athletes with highly accurate and timely feedback. Through the utilization of multi-modal robot data, CAM-Vtrans offers valuable assistance, enabling athletes to optimize their performance while minimizing potential injury risks. This novel approach represents a significant advancement in the field, offering an innovative solution to overcome the limitations of existing methods and enhance the precision and efficiency of sports training programs.

摘要

引言

辅助机器人和人机交互已成为体育训练的重要组成部分。然而,现有方法往往无法提供实时且准确的反馈,并且常常缺乏综合多模态数据的整合。

方法

为了解决这些问题,我们提出了一种开创性的创新方法:CAM-Vtrans-交叉注意力多模态视觉变换器。通过利用诸如视觉变换器(ViT)等先进技术以及CLIP等模型的优势,结合交叉注意力机制,CAM-Vtrans利用视觉和文本信息的力量为运动员提供高度准确和及时的反馈。通过利用多模态机器人数据,CAM-Vtrans提供了有价值的帮助,使运动员能够优化其表现,同时将潜在的受伤风险降至最低。这种新颖的方法代表了该领域的重大进步,提供了一种创新的解决方案,以克服现有方法的局限性,并提高体育训练计划的精度和效率。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/208c/11502466/42921a4d5369/fnbot-18-1453571-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/208c/11502466/e047019e043c/fnbot-18-1453571-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/208c/11502466/df2f7d11cfa2/fnbot-18-1453571-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/208c/11502466/49afdd2e3d62/fnbot-18-1453571-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/208c/11502466/c3c30f326986/fnbot-18-1453571-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/208c/11502466/42921a4d5369/fnbot-18-1453571-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/208c/11502466/e047019e043c/fnbot-18-1453571-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/208c/11502466/df2f7d11cfa2/fnbot-18-1453571-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/208c/11502466/49afdd2e3d62/fnbot-18-1453571-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/208c/11502466/c3c30f326986/fnbot-18-1453571-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/208c/11502466/42921a4d5369/fnbot-18-1453571-g0005.jpg

相似文献

1
CAM-Vtrans: real-time sports training utilizing multi-modal robot data.CAM-Vtrans:利用多模态机器人数据的实时运动训练
Front Neurorobot. 2024 Oct 11;18:1453571. doi: 10.3389/fnbot.2024.1453571. eCollection 2024.
2
Sports competition tactical analysis model of cross-modal transfer learning intelligent robot based on Swin Transformer and CLIP.基于Swin Transformer和CLIP的跨模态迁移学习智能机器人体育竞赛战术分析模型
Front Neurorobot. 2023 Oct 30;17:1275645. doi: 10.3389/fnbot.2023.1275645. eCollection 2023.
3
Cross-modal self-attention mechanism for controlling robot volleyball motion.用于控制机器人排球运动的跨模态自注意力机制。
Front Neurorobot. 2023 Nov 10;17:1288463. doi: 10.3389/fnbot.2023.1288463. eCollection 2023.
4
Res-FLNet: human-robot interaction and collaboration for multi-modal sensing robot autonomous driving tasks based on learning control algorithm.Res-FLNet:基于学习控制算法的用于多模态传感机器人自动驾驶任务的人机交互与协作
Front Neurorobot. 2023 Oct 2;17:1269105. doi: 10.3389/fnbot.2023.1269105. eCollection 2023.
5
A modality-collaborative convolution and transformer hybrid network for unpaired multi-modal medical image segmentation with limited annotations.一种用于具有有限标注的未配对多模态医学图像分割的模态协作卷积与Transformer混合网络。
Med Phys. 2023 Sep;50(9):5460-5478. doi: 10.1002/mp.16338. Epub 2023 Mar 15.
6
Research on deep reinforcement learning basketball robot shooting skills improvement based on end to end architecture and multi-modal perception.基于端到端架构和多模态感知的深度强化学习篮球机器人投篮技术改进研究
Front Neurorobot. 2023 Oct 13;17:1274543. doi: 10.3389/fnbot.2023.1274543. eCollection 2023.
7
SwinCross: Cross-modal Swin transformer for head-and-neck tumor segmentation in PET/CT images.SwinCross:用于 PET/CT 图像中头颈部肿瘤分割的跨模态 Swin 变换器。
Med Phys. 2024 Mar;51(3):2096-2107. doi: 10.1002/mp.16703. Epub 2023 Sep 30.
8
What Does a Language-And-Vision Transformer See: The Impact of Semantic Information on Visual Representations.语言与视觉Transformer看到了什么:语义信息对视觉表征的影响。
Front Artif Intell. 2021 Dec 3;4:767971. doi: 10.3389/frai.2021.767971. eCollection 2021.
9
TransVG++: End-to-End Visual Grounding With Language Conditioned Vision Transformer.TransVG++:基于语言条件视觉Transformer的端到端视觉基础
IEEE Trans Pattern Anal Mach Intell. 2023 Nov;45(11):13636-13652. doi: 10.1109/TPAMI.2023.3296823. Epub 2023 Oct 3.
10
A 3D hierarchical cross-modality interaction network using transformers and convolutions for brain glioma segmentation in MR images.一种使用变换和卷积的 3D 层次跨模态交互网络,用于磁共振图像中的脑胶质瘤分割。
Med Phys. 2024 Nov;51(11):8371-8389. doi: 10.1002/mp.17354. Epub 2024 Aug 13.

引用本文的文献

1
An improved graph factorization machine based on solving unbalanced game perception.一种基于解决不平衡博弈感知的改进型图分解机
Front Neurorobot. 2024 Dec 4;18:1481297. doi: 10.3389/fnbot.2024.1481297. eCollection 2024.

本文引用的文献

1
A Multi-Modal Egocentric Activity Recognition Approach towards Video Domain Generalization.一种面向视频领域泛化的多模态自我中心活动识别方法。
Sensors (Basel). 2024 Apr 12;24(8):2491. doi: 10.3390/s24082491.
2
Botulinum toxin treatment may improve myoelectric pattern recognition in robot-assisted stroke rehabilitation.肉毒杆菌毒素治疗可能会改善机器人辅助中风康复中的肌电模式识别。
Front Neurosci. 2024 Feb 29;18:1364214. doi: 10.3389/fnins.2024.1364214. eCollection 2024.
3
Exploring wireless device-free localization technique to assist home-based neuro-rehabilitation.
探索无设备无线定位技术以辅助居家神经康复。
Front Neurosci. 2024 Feb 2;18:1344841. doi: 10.3389/fnins.2024.1344841. eCollection 2024.
4
EEG generation mechanism of lower limb active movement intention and its virtual reality induction enhancement: a preliminary study.下肢主动运动意图的脑电图生成机制及其虚拟现实诱导增强:一项初步研究。
Front Neurosci. 2024 Jan 30;17:1305850. doi: 10.3389/fnins.2023.1305850. eCollection 2023.
5
Cross-modal self-attention mechanism for controlling robot volleyball motion.用于控制机器人排球运动的跨模态自注意力机制。
Front Neurorobot. 2023 Nov 10;17:1288463. doi: 10.3389/fnbot.2023.1288463. eCollection 2023.
6
Wearable Sensor-Based Human Activity Recognition with Transformer Model.基于可穿戴传感器的Transformer 模型人体活动识别。
Sensors (Basel). 2022 Mar 1;22(5):1911. doi: 10.3390/s22051911.
7
A Novel Robot-Aided Upper Limb Rehabilitation Training System Based on Multimodal Feedback.一种基于多模态反馈的新型机器人辅助上肢康复训练系统。
Front Robot AI. 2019 Nov 8;6:102. doi: 10.3389/frobt.2019.00102. eCollection 2019.
8
Skill training in multimodal virtual environments.多模态虚拟环境中的技能训练。
Work. 2012;41 Suppl 1:2284-7. doi: 10.3233/WOR-2012-0452-2284.