• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

课堂视频中的双模态学习参与度识别。

Bimodal Learning Engagement Recognition from Videos in the Classroom.

机构信息

Hubei Research Center for Educational Informationization, Faculty of Artificial Intelligence in Education, Central China Normal University, Wuhan 430074, China.

Huanggang High School of Hubei Province, Huanggang 438000, China.

出版信息

Sensors (Basel). 2022 Aug 9;22(16):5932. doi: 10.3390/s22165932.

DOI:10.3390/s22165932
PMID:36015693
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9415674/
Abstract

Engagement plays an essential role in the learning process. Recognition of learning engagement in the classroom helps us understand the student's learning state and optimize the teaching and study processes. Traditional recognition methods such as self-report and teacher observation are time-consuming and obtrusive to satisfy the needs of large-scale classrooms. With the development of big data analysis and artificial intelligence, applying intelligent methods such as deep learning to recognize learning engagement has become the research hotspot in education. In this paper, based on non-invasive classroom videos, first, a multi-cues classroom learning engagement database was constructed. Then, we introduced the power IoU loss function to You Only Look Once version 5 (YOLOv5) to detect the students and obtained a precision of 95.4%. Finally, we designed a bimodal learning engagement recognition method based on ResNet50 and CoAtNet. Our proposed bimodal learning engagement method obtained an accuracy of 93.94% using the KNN classifier. The experimental results confirmed that the proposed method outperforms most state-of-the-art techniques.

摘要

参与在学习过程中起着至关重要的作用。在课堂上识别学习参与度有助于我们了解学生的学习状态,并优化教学和学习过程。传统的识别方法,如自我报告和教师观察,既耗时又繁琐,难以满足大规模课堂的需求。随着大数据分析和人工智能的发展,应用深度学习等智能方法来识别学习参与度已成为教育领域的研究热点。在本文中,我们基于非侵入式课堂视频,首先构建了一个多线索课堂学习参与度数据库。然后,我们将 Power IoU 损失函数引入到 You Only Look Once 版本 5(YOLOv5)中,以检测学生,得到了 95.4%的精度。最后,我们设计了一种基于 ResNet50 和 CoAtNet 的双模态学习参与度识别方法。我们提出的双模态学习参与度方法使用 KNN 分类器获得了 93.94%的准确率。实验结果证实了该方法优于大多数最先进的技术。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/92a8c71b4cc0/sensors-22-05932-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/8af90fb565ce/sensors-22-05932-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/9e61ed38e9be/sensors-22-05932-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/e3310a92a98d/sensors-22-05932-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/6e5042c48925/sensors-22-05932-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/e035c630ee21/sensors-22-05932-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/c08f314d09ea/sensors-22-05932-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/789fda0fa894/sensors-22-05932-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/5c58133f799a/sensors-22-05932-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/2dfb1b3c7145/sensors-22-05932-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/5ead3aad1d9a/sensors-22-05932-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/dacc3d813e02/sensors-22-05932-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/c5e70351d354/sensors-22-05932-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/92a8c71b4cc0/sensors-22-05932-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/8af90fb565ce/sensors-22-05932-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/9e61ed38e9be/sensors-22-05932-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/e3310a92a98d/sensors-22-05932-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/6e5042c48925/sensors-22-05932-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/e035c630ee21/sensors-22-05932-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/c08f314d09ea/sensors-22-05932-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/789fda0fa894/sensors-22-05932-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/5c58133f799a/sensors-22-05932-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/2dfb1b3c7145/sensors-22-05932-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/5ead3aad1d9a/sensors-22-05932-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/dacc3d813e02/sensors-22-05932-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/c5e70351d354/sensors-22-05932-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/473e/9415674/92a8c71b4cc0/sensors-22-05932-g013.jpg

相似文献

1
Bimodal Learning Engagement Recognition from Videos in the Classroom.课堂视频中的双模态学习参与度识别。
Sensors (Basel). 2022 Aug 9;22(16):5932. doi: 10.3390/s22165932.
2
Automated Student Classroom Behaviors' Perception and Identification Using Motion Sensors.利用运动传感器对学生课堂行为进行自动感知与识别
Bioengineering (Basel). 2023 Jan 18;10(2):127. doi: 10.3390/bioengineering10020127.
3
Analysis of the Current Situation of Teaching and Learning of Ideological and Political Theory Courses by Deep Learning.深度学习视域下的思想政治理论课教与学现状分析。
Comput Intell Neurosci. 2022 Aug 18;2022:5396054. doi: 10.1155/2022/5396054. eCollection 2022.
4
Using the Flipped Classroom to Bridge the Gap to Generation Y.利用翻转课堂弥合与Y世代的差距。
Ochsner J. 2016 Spring;16(1):32-6.
5
Flipped classrooms and student learning: not just surface gains.翻转课堂与学生学习:不止于表面收获。
Adv Physiol Educ. 2016 Mar;40(1):47-55. doi: 10.1152/advan.00098.2015.
6
Comparison of team-based learning vs. lecture-based teaching with small group discussion in a master's degree in nursing education course.团队学习与基于讲座的教学与小组讨论在护理教育硕士课程中的比较。
Nurse Educ Today. 2021 Oct;105:105043. doi: 10.1016/j.nedt.2021.105043. Epub 2021 Jul 2.
7
Learning Preferences and Engagement Level of Generation Z Nursing Students.Z 世代护理专业学生的学习偏好和参与度。
Nurse Educ. 2020 May/Jun;45(3):160-164. doi: 10.1097/NNE.0000000000000710.
8
Design and Implementation of IoT Data-Driven Intelligent Law Classroom Teaching System.物联网数据驱动的智能法律课堂教学系统的设计与实现。
Comput Intell Neurosci. 2022 Mar 25;2022:8003909. doi: 10.1155/2022/8003909. eCollection 2022.
9
Comparing the effects of traditional lecture and flipped classroom on nursing students' critical thinking disposition: A quasi-experimental study.比较传统讲座和翻转课堂对护生批判性思维倾向的影响:一项准实验研究。
Nurse Educ Today. 2018 Dec;71:151-156. doi: 10.1016/j.nedt.2018.09.027. Epub 2018 Sep 28.
10
Design of a Smart Art Classroom System Based on Internet of Things.基于物联网的智能美术教室系统设计。
Comput Intell Neurosci. 2022 Apr 12;2022:9257827. doi: 10.1155/2022/9257827. eCollection 2022.

引用本文的文献

1
DMSA-Net: a deformable multiscale adaptive classroom behavior recognition network.DMSA-Net:一种可变形多尺度自适应课堂行为识别网络。
PeerJ Comput Sci. 2025 Apr 30;11:e2876. doi: 10.7717/peerj-cs.2876. eCollection 2025.
2
CSSA-YOLO: Cross-Scale Spatiotemporal Attention Network for Fine-Grained Behavior Recognition in Classroom Environments.CSSA-YOLO:用于课堂环境中细粒度行为识别的跨尺度时空注意力网络。
Sensors (Basel). 2025 May 15;25(10):3132. doi: 10.3390/s25103132.
3
Classroom Behavior Recognition Using Computer Vision: A Systematic Review.
基于计算机视觉的课堂行为识别:系统综述
Sensors (Basel). 2025 Jan 10;25(2):373. doi: 10.3390/s25020373.
4
Smart Classrooms: How Sensors and AI Are Shaping Educational Paradigms.智能教室:传感器和人工智能如何改变教育范式。
Sensors (Basel). 2024 Aug 24;24(17):5487. doi: 10.3390/s24175487.
5
Automated Monitoring of Human-Computer Interaction for Assessing Teachers' Digital Competence Based on LMS Data Extraction.基于学习管理系统(LMS)数据提取的人机交互自动监测以评估教师的数字能力
Sensors (Basel). 2024 May 23;24(11):3326. doi: 10.3390/s24113326.
6
Student Behavior Detection in the Classroom Based on Improved YOLOv8.基于改进YOLOv8的课堂学生行为检测
Sensors (Basel). 2023 Oct 11;23(20):8385. doi: 10.3390/s23208385.
7
MSTA-SlowFast: A Student Behavior Detector for Classroom Environments.MSTA-SlowFast:一种用于课堂环境的学生行为检测器。
Sensors (Basel). 2023 May 30;23(11):5205. doi: 10.3390/s23115205.