Hossain Delwar, Ghosh Tonmoy, Sazonov Edward
Electrical and Computer Engineering Department, The University of Alabama, Tuscaloosa, AL 35487, USA.
IEEE Access. 2020;8:101934-101945. doi: 10.1109/access.2020.2998716. Epub 2020 Jun 1.
Methods for measuring of eating behavior (known as meal microstructure) often rely on manual annotation of bites, chews, and swallows on meal videos or wearable sensor signals. The manual annotation may be time consuming and erroneous, while wearable sensors may not capture every aspect of eating (e.g. chews only). The aim of this study is to develop a method to detect and count bites and chews automatically from meal videos. The method was developed on a dataset of 28 volunteers consuming unrestricted meals in the laboratory under video observation. First, the faces in the video (regions of interest, ROI) were detected using Faster R-CNN. Second, a pre-trained AlexNet was trained on the detected faces to classify images as a bite/no bite image. Third, the affine optical flow was applied in consecutively detected faces to find the rotational movement of the pixels in the ROIs. The number of chews in a meal video was counted by converting the 2-D images to a 1-D optical flow parameter and finding peaks. The developed bite and chew count algorithm was applied to 84 meal videos collected from 28 volunteers. A mean accuracy (±STD) of 85.4% (±6.3%) with respect to manual annotation was obtained for the number of bites and 88.9% (±7.4%) for the number of chews. The proposed method for an automatic bite and chew counting shows promising results that can be used as an alternative solution to manual annotation.
测量进食行为(即进餐微观结构)的方法通常依赖于对进餐视频或可穿戴传感器信号中的咬、嚼和吞咽动作进行人工标注。人工标注可能既耗时又容易出错,而可穿戴传感器可能无法捕捉到进食的各个方面(例如只能捕捉嚼的动作)。本研究的目的是开发一种从进餐视频中自动检测和计数咬和嚼动作的方法。该方法是在一个包含28名志愿者在实验室视频观察下自由进餐的数据集上开发的。首先,使用更快的区域卷积神经网络(Faster R-CNN)检测视频中的人脸(感兴趣区域,ROI)。其次,在检测到的人脸上训练一个预训练的亚历克斯网络(AlexNet),将图像分类为咬/未咬图像。第三,将仿射光流应用于连续检测到的人脸,以找到感兴趣区域中像素的旋转运动。通过将二维图像转换为一维光流参数并找到峰值来计算进餐视频中的嚼的次数。将开发的咬和嚼计数算法应用于从28名志愿者收集的84个进餐视频。对于咬的次数,相对于人工标注获得的平均准确率(±标准差)为85.4%(±6.3%),对于嚼的次数为88.9%(±7.4%)。所提出的自动咬和嚼计数方法显示出有前景的结果,可作为人工标注的替代解决方案。