• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于 3D 卷积神经网络的帧序列识别的单目摄像机面部位置不变驾驶员注视区域分类器。

Single Camera Face Position-Invariant Driver's Gaze Zone Classifier Based on Frame-Sequence Recognition Using 3D Convolutional Neural Networks.

机构信息

Graduate School of Creative Science and Engineering, Waseda University, Tokyo 169-8555, Japan.

Research Institute for Science and Engineering (RISE), Waseda University, Tokyo 162-0044, Japan.

出版信息

Sensors (Basel). 2022 Aug 5;22(15):5857. doi: 10.3390/s22155857.

DOI:10.3390/s22155857
PMID:35957412
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9370862/
Abstract

Estimating the driver's gaze in a natural real-world setting can be problematic for different challenging scenario conditions. For example, faces will undergo facial occlusions, illumination, or various face positions while driving. In this effort, we aim to reduce misclassifications in driving situations when the driver has different face distances regarding the camera. Three-dimensional Convolutional Neural Networks (CNN) models can make a spatio-temporal driver's representation that extracts features encoded in multiple adjacent frames that can describe motions. This characteristic may help ease the deficiencies of a per-frame recognition system due to the lack of context information. For example, the front, navigator, right window, left window, back mirror, and speed meter are part of the known common areas to be checked by drivers. Based on this, we implement and evaluate a model that is able to detect the head direction toward these regions having various distances from the camera. In our evaluation, the 2D CNN model had a mean average recall of 74.96% across the three models, whereas the 3D CNN model had a mean average recall of 87.02%. This result show that our proposed 3D CNN-based approach outperforms a 2D CNN per-frame recognition approach in driving situations when the driver's face has different distances from the camera.

摘要

在自然真实环境中估计驾驶员的注视方向对于不同的挑战性场景条件可能是一个问题。例如,在驾驶过程中,人脸会经历面部遮挡、光照或各种面部位置。在这项工作中,我们旨在减少驾驶员与摄像机的面部距离不同时驾驶情况下的误分类。三维卷积神经网络(CNN)模型可以构建驾驶员的时空表示,提取多个相邻帧中编码的特征,这些特征可以描述运动。由于缺乏上下文信息,这一特征可能有助于缓解基于单帧识别系统的不足。例如,前窗、导航、右窗、左窗、后镜和速度计是驾驶员需要检查的已知常见区域的一部分。基于此,我们实现并评估了一种能够检测头部朝向这些区域的模型,这些区域与摄像机的距离各不相同。在我们的评估中,2D CNN 模型在三个模型中的平均召回率为 74.96%,而 3D CNN 模型的平均召回率为 87.02%。这一结果表明,与基于单帧识别的 2D CNN 方法相比,我们提出的基于 3D CNN 的方法在驾驶员与摄像机的面部距离不同的驾驶情况下表现更好。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/78f88d30223e/sensors-22-05857-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/9e9579b796a1/sensors-22-05857-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/270f6ebfda5c/sensors-22-05857-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/561f17d2544e/sensors-22-05857-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/978d0fec8682/sensors-22-05857-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/5be1fd617fa4/sensors-22-05857-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/a1adfa40ebf6/sensors-22-05857-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/d6d643067401/sensors-22-05857-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/72391a48054f/sensors-22-05857-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/78f88d30223e/sensors-22-05857-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/9e9579b796a1/sensors-22-05857-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/270f6ebfda5c/sensors-22-05857-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/561f17d2544e/sensors-22-05857-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/978d0fec8682/sensors-22-05857-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/5be1fd617fa4/sensors-22-05857-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/a1adfa40ebf6/sensors-22-05857-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/d6d643067401/sensors-22-05857-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/72391a48054f/sensors-22-05857-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c244/9370862/78f88d30223e/sensors-22-05857-g009.jpg

相似文献

1
Single Camera Face Position-Invariant Driver's Gaze Zone Classifier Based on Frame-Sequence Recognition Using 3D Convolutional Neural Networks.基于 3D 卷积神经网络的帧序列识别的单目摄像机面部位置不变驾驶员注视区域分类器。
Sensors (Basel). 2022 Aug 5;22(15):5857. doi: 10.3390/s22155857.
2
Faster R-CNN and Geometric Transformation-Based Detection of Driver's Eyes Using Multiple Near-Infrared Camera Sensors.基于 Faster R-CNN 和几何变换的使用多个近红外相机传感器的驾驶员眼睛检测。
Sensors (Basel). 2019 Jan 7;19(1):197. doi: 10.3390/s19010197.
3
DRER: Deep Learning-Based Driver's Real Emotion Recognizer.DRER:基于深度学习的驾驶员真实情感识别器。
Sensors (Basel). 2021 Mar 19;21(6):2166. doi: 10.3390/s21062166.
4
A Driver's Visual Attention Prediction Using Optical Flow.基于光流的驾驶员视觉注意力预测
Sensors (Basel). 2021 May 27;21(11):3722. doi: 10.3390/s21113722.
5
Research on imaging method of driver's attention area based on deep neural network.基于深度神经网络的驾驶员注意区域成像方法研究。
Sci Rep. 2022 Sep 30;12(1):16427. doi: 10.1038/s41598-022-20829-w.
6
Driver's Head Pose and Gaze Zone Estimation Based on Multi-Zone Templates Registration and Multi-Frame Point Cloud Fusion.基于多区域模板配准和多帧点云融合的驾驶员头部姿势和注视区域估计。
Sensors (Basel). 2022 Apr 20;22(9):3154. doi: 10.3390/s22093154.
7
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors.基于卷积神经网络的多模态摄像头传感器对激进驾驶和平稳驾驶过程中驾驶员情绪的分类
Sensors (Basel). 2018 Mar 23;18(4):957. doi: 10.3390/s18040957.
8
A Hybrid Model for Driver Emotion Detection Using Feature Fusion Approach.基于特征融合方法的驾驶员情绪检测混合模型。
Int J Environ Res Public Health. 2022 Mar 6;19(5):3085. doi: 10.3390/ijerph19053085.
9
A Driver Gaze Estimation Method Based on Deep Learning.基于深度学习的驾驶员注视估计方法。
Sensors (Basel). 2022 May 23;22(10):3959. doi: 10.3390/s22103959.
10
Driving event recognition using machine learning and smartphones.使用机器学习和智能手机进行驾驶行为识别。
F1000Res. 2022 Jan 18;11:57. doi: 10.12688/f1000research.73134.2. eCollection 2022.

本文引用的文献

1
Catastrophic factors involved in road accidents: Underlying causes and descriptive analysis.道路交通事故中的灾难性因素:根本原因和描述性分析。
PLoS One. 2019 Oct 9;14(10):e0223473. doi: 10.1371/journal.pone.0223473. eCollection 2019.
2
A Comprehensive Survey of Driving Monitoring and Assistance Systems.驾驶监测与辅助系统综述
Sensors (Basel). 2019 Jun 6;19(11):2574. doi: 10.3390/s19112574.
3
Continuous Driver's Gaze Zone Estimation Using RGB-D Camera.基于 RGB-D 相机的驾驶员注视区域连续估计
Sensors (Basel). 2019 Mar 14;19(6):1287. doi: 10.3390/s19061287.
4
Deep Learning-Based Gaze Detection System for Automobile Drivers Using a NIR Camera Sensor.基于深度学习的汽车驾驶员凝视检测系统,使用近红外相机传感器。
Sensors (Basel). 2018 Feb 3;18(2):456. doi: 10.3390/s18020456.
5
Deep learning.深度学习。
Nature. 2015 May 28;521(7553):436-44. doi: 10.1038/nature14539.
6
Reducing the dimensionality of data with neural networks.使用神经网络降低数据维度。
Science. 2006 Jul 28;313(5786):504-7. doi: 10.1126/science.1127647.