• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于深度视频的车辆二次动作识别:卷积神经网络和具有空间增强注意力机制的双向长短时记忆

Depth Video-Based Secondary Action Recognition in Vehicles via Convolutional Neural Network and Bidirectional Long Short-Term Memory with Spatial Enhanced Attention Mechanism.

机构信息

Graduate School of Science and Technology, Keio University, Yokohama 223-8522, Japan.

Faculty of Science and Technology, Keio University, Yokohama 223-8522, Japan.

出版信息

Sensors (Basel). 2024 Oct 13;24(20):6604. doi: 10.3390/s24206604.

DOI:10.3390/s24206604
PMID:39460085
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11510851/
Abstract

Secondary actions in vehicles are activities that drivers engage in while driving that are not directly related to the primary task of operating the vehicle. Secondary Action Recognition (SAR) in drivers is vital for enhancing road safety and minimizing accidents related to distracted driving. It also plays an important part in modern car driving systems such as Advanced Driving Assistance Systems (ADASs), as it helps identify distractions and predict the driver's intent. Traditional methods of action recognition in vehicles mostly rely on RGB videos, which can be significantly impacted by external conditions such as low light levels. In this research, we introduce a novel method for SAR. Our approach utilizes depth-video data obtained from a depth sensor located in a vehicle. Our methodology leverages the Convolutional Neural Network (CNN), which is enhanced by the Spatial Enhanced Attention Mechanism (SEAM) and combined with Bidirectional Long Short-Term Memory (Bi-LSTM) networks. This method significantly enhances action recognition ability in depth videos by improving both the spatial and temporal aspects. We conduct experiments using K-fold cross validation, and the experimental results show that on the public benchmark dataset Drive&Act, our proposed method shows significant improvement in SAR compared to the state-of-the-art methods, reaching an accuracy of about 84% in SAR in depth videos.

摘要

车辆中的次要动作是驾驶员在驾驶时进行的与操作车辆的主要任务不直接相关的活动。驾驶员的次要动作识别(SAR)对于提高道路安全和最大限度地减少与分心驾驶相关的事故至关重要。它在现代驾驶系统中也起着重要作用,如高级驾驶辅助系统(ADAS),因为它有助于识别分心并预测驾驶员的意图。车辆中传统的动作识别方法主要依赖于 RGB 视频,而这些视频可能会受到外部条件的显著影响,如低光照水平。在这项研究中,我们引入了一种新的 SAR 方法。我们的方法利用了从车辆中安装的深度传感器获得的深度视频数据。我们的方法利用了卷积神经网络(CNN),通过空间增强注意力机制(SEAM)进行增强,并结合了双向长短期记忆(Bi-LSTM)网络。这种方法通过提高空间和时间方面的性能,显著提高了深度视频中的动作识别能力。我们使用 K 折交叉验证进行实验,实验结果表明,在公共基准数据集 Drive&Act 上,与最先进的方法相比,我们提出的方法在 SAR 方面有显著的改进,在深度视频中的 SAR 准确性达到了约 84%。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/617ffc02397f/sensors-24-06604-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/87bed4e35d12/sensors-24-06604-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/4871d09b0560/sensors-24-06604-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/15e05169195e/sensors-24-06604-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/b931f477c484/sensors-24-06604-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/d27e6a911716/sensors-24-06604-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/da99a183e34a/sensors-24-06604-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/fa0726a02c7b/sensors-24-06604-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/e822342422be/sensors-24-06604-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/5117716eb05c/sensors-24-06604-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/194e2bb5d5e3/sensors-24-06604-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/544d0d373679/sensors-24-06604-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/ddf8b75addcd/sensors-24-06604-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/472233ae4b00/sensors-24-06604-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/cffa6730e981/sensors-24-06604-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/5d861e835001/sensors-24-06604-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/617ffc02397f/sensors-24-06604-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/87bed4e35d12/sensors-24-06604-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/4871d09b0560/sensors-24-06604-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/15e05169195e/sensors-24-06604-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/b931f477c484/sensors-24-06604-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/d27e6a911716/sensors-24-06604-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/da99a183e34a/sensors-24-06604-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/fa0726a02c7b/sensors-24-06604-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/e822342422be/sensors-24-06604-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/5117716eb05c/sensors-24-06604-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/194e2bb5d5e3/sensors-24-06604-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/544d0d373679/sensors-24-06604-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/ddf8b75addcd/sensors-24-06604-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/472233ae4b00/sensors-24-06604-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/cffa6730e981/sensors-24-06604-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/5d861e835001/sensors-24-06604-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4596/11510851/617ffc02397f/sensors-24-06604-g016.jpg

相似文献

1
Depth Video-Based Secondary Action Recognition in Vehicles via Convolutional Neural Network and Bidirectional Long Short-Term Memory with Spatial Enhanced Attention Mechanism.基于深度视频的车辆二次动作识别:卷积神经网络和具有空间增强注意力机制的双向长短时记忆
Sensors (Basel). 2024 Oct 13;24(20):6604. doi: 10.3390/s24206604.
2
A Hybrid Deep Learning Model for Recognizing Actions of Distracted Drivers.用于识别分心驾驶员行为的混合深度学习模型。
Sensors (Basel). 2021 Nov 8;21(21):7424. doi: 10.3390/s21217424.
3
CBAM VGG16: An efficient driver distraction classification using CBAM embedded VGG16 architecture.CBAM-VGG16:一种使用嵌入 CBAM 的 VGG16 架构的高效驾驶员分心分类方法。
Comput Biol Med. 2024 Sep;180:108945. doi: 10.1016/j.compbiomed.2024.108945. Epub 2024 Aug 1.
4
Research on a Cognitive Distraction Recognition Model for Intelligent Driving Systems Based on Real Vehicle Experiments.基于实车实验的智能驾驶系统认知分心识别模型研究。
Sensors (Basel). 2020 Aug 7;20(16):4426. doi: 10.3390/s20164426.
5
Sensor-Based Classification of Primary and Secondary Car Driver Activities Using Convolutional Neural Networks.基于卷积神经网络的主、副驾驶行为的传感器分类。
Sensors (Basel). 2023 Jun 13;23(12):5551. doi: 10.3390/s23125551.
6
Brain tumor segmentation and detection in MRI using convolutional neural networks and VGG16.使用卷积神经网络和VGG16在磁共振成像(MRI)中进行脑肿瘤分割与检测
Cancer Biomark. 2025 Mar;42(3):18758592241311184. doi: 10.1177/18758592241311184. Epub 2025 Apr 4.
7
Spatial-frequency-temporal convolutional recurrent network for olfactory-enhanced EEG emotion recognition.基于空间频率-时间卷积循环网络的嗅觉增强脑电情感识别
J Neurosci Methods. 2022 Jul 1;376:109624. doi: 10.1016/j.jneumeth.2022.109624. Epub 2022 May 16.
8
Real-Time Driving Distraction Recognition Through a Wrist-Mounted Accelerometer.基于腕部佩戴加速度计的实时驾驶分心识别。
Hum Factors. 2022 Dec;64(8):1412-1428. doi: 10.1177/0018720821995000. Epub 2021 Feb 24.
9
Research on imaging method of driver's attention area based on deep neural network.基于深度神经网络的驾驶员注意区域成像方法研究。
Sci Rep. 2022 Sep 30;12(1):16427. doi: 10.1038/s41598-022-20829-w.
10
Trajectory-level fog detection based on in-vehicle video camera with TensorFlow deep learning utilizing SHRP2 naturalistic driving data.基于 TensorFlow 深度学习利用 SHRP2 自然驾驶数据的车载视频摄像机的轨迹级雾检测。
Accid Anal Prev. 2020 Jul;142:105521. doi: 10.1016/j.aap.2020.105521. Epub 2020 May 11.

本文引用的文献

1
Hybrid convolution neural network with channel attention mechanism for sensor-based human activity recognition.基于传感器的人体活动识别的带通道注意力机制的混合卷积神经网络。
Sci Rep. 2023 Jul 26;13(1):12067. doi: 10.1038/s41598-023-39080-y.
2
Human activity recognition using tools of convolutional neural networks: A state of the art review, data sets, challenges, and future prospects.使用卷积神经网络工具进行人类活动识别:最新研究综述、数据集、挑战与未来展望。
Comput Biol Med. 2022 Oct;149:106060. doi: 10.1016/j.compbiomed.2022.106060. Epub 2022 Sep 1.
3
Human Action Recognition From Various Data Modalities: A Review.
基于多种数据模态的人类行为识别综述
IEEE Trans Pattern Anal Mach Intell. 2023 Mar;45(3):3200-3225. doi: 10.1109/TPAMI.2022.3183112. Epub 2023 Feb 3.
4
MMNet: A Model-Based Multimodal Network for Human Action Recognition in RGB-D Videos.MMNet:一种基于模型的 RGB-D 视频人体动作识别多模态网络。
IEEE Trans Pattern Anal Mach Intell. 2023 Mar;45(3):3522-3538. doi: 10.1109/TPAMI.2022.3177813. Epub 2023 Feb 3.
5
A union of deep learning and swarm-based optimization for 3D human action recognition.基于深度学习和群体智能优化的三维人体动作识别方法。
Sci Rep. 2022 Mar 31;12(1):5494. doi: 10.1038/s41598-022-09293-8.
6
Indirect Time-of-Flight Depth Sensor with Two-Step Comparison Scheme for Depth Frame Difference Detection.用于深度帧差检测的具有两步比较方案的间接飞行时间深度传感器。
Sensors (Basel). 2019 Aug 23;19(17):3674. doi: 10.3390/s19173674.
7
Long-Term Recurrent Convolutional Networks for Visual Recognition and Description.长期递归卷积网络的视觉识别与描述。
IEEE Trans Pattern Anal Mach Intell. 2017 Apr;39(4):677-691. doi: 10.1109/TPAMI.2016.2599174. Epub 2016 Sep 1.