Bhavya Bhavya, Chen Si, Zhang Zhilin, Li Wenting, Zhai Chengxiang, Angrave Lawrence, Huang Yun
Department of Computer Science, University of Illinois at Urbana-Champaign, Urbana, IL 61801 USA.
School of Information Sciences, University of Illinois at Urbana-Champaign, Champaign, IL 61820 USA.
Educ Technol Res Dev. 2022;70(5):1755-1779. doi: 10.1007/s11423-022-10137-5. Epub 2022 Jul 15.
Captions play a major role in making educational videos accessible to all and are known to benefit a wide range of learners. However, many educational videos either do not have captions or have inaccurate captions. Prior work has shown the benefits of using crowdsourcing to obtain accurate captions in a cost-efficient way, though there is a lack of understanding of how learners edit captions of educational videos either individually or collaboratively. In this work, we conducted a user study where 58 learners (in a course of 387 learners) participated in the editing of captions in 89 lecture videos that were generated by Automatic Speech Recognition (ASR) technologies. For each video, different learners conducted two rounds of editing. Based on editing logs, we created a taxonomy of errors in educational video captions (e.g., Discipline-Specific, General, Equations). From the interviews, we identified individual and collaborative error editing strategies. We then further demonstrated the feasibility of applying machine learning models to assist learners in editing. Our work provides practical implications for advancing video-based learning and for educational video caption editing.
字幕在使教育视频能够被所有人访问方面发挥着重要作用,并且已知对广泛的学习者有益。然而,许多教育视频要么没有字幕,要么字幕不准确。先前的工作已经表明利用众包以经济高效的方式获得准确字幕的好处,尽管对于学习者如何单独或协作编辑教育视频的字幕缺乏了解。在这项工作中,我们进行了一项用户研究,其中58名学习者(在一门有387名学习者的课程中)参与了由自动语音识别(ASR)技术生成的89个讲座视频的字幕编辑。对于每个视频,不同的学习者进行了两轮编辑。基于编辑日志,我们创建了教育视频字幕错误的分类法(例如,特定学科、一般、方程式)。通过访谈,我们确定了个人和协作错误编辑策略。然后,我们进一步证明了应用机器学习模型协助学习者进行编辑的可行性。我们的工作为推进基于视频的学习和教育视频字幕编辑提供了实际意义。