Haberfehlner Helga, van de Ven Shankara S, van der Burg Sven A, Huber Florian, Georgievska Sonja, Aleo Ignazio, Harlaar Jaap, Bonouvrié Laura A, van der Krogt Marjolein M, Buizer Annemieke I
Amsterdam UMC location Vrije Universiteit Amsterdam, Rehabilitation Medicine, Amsterdam, Netherlands.
Amsterdam Movement Sciences, Rehabilitation and Development, Amsterdam, Netherlands.
Front Robot AI. 2023 Mar 2;10:1108114. doi: 10.3389/frobt.2023.1108114. eCollection 2023.
Video-based clinical rating plays an important role in assessing dystonia and monitoring the effect of treatment in dyskinetic cerebral palsy (CP). However, evaluation by clinicians is time-consuming, and the quality of rating is dependent on experience. The aim of the current study is to provide a proof-of-concept for a machine learning approach to automatically assess scoring of dystonia using 2D stick figures extracted from videos. Model performance was compared to human performance. A total of 187 video sequences of 34 individuals with dyskinetic CP (8-23 years, all non-ambulatory) were filmed at rest during lying and supported sitting. Videos were scored by three raters according to the Dyskinesia Impairment Scale (DIS) for arm and leg dystonia (normalized scores ranging from 0-1). Coordinates in pixels of the left and right wrist, elbow, shoulder, hip, knee and ankle were extracted using DeepLabCut, an open source toolbox that builds on a pose estimation algorithm. Within a subset, tracking accuracy was assessed for a pretrained human model and for models trained with an increasing number of manually labeled frames. The mean absolute error (MAE) between DeepLabCut's prediction of the position of body points and manual labels was calculated. Subsequently, movement and position features were calculated from extracted body point coordinates. These features were fed into a Random Forest Regressor to train a model to predict the clinical scores. The model performance trained with data from one rater evaluated by MAEs (model-rater) was compared to inter-rater accuracy. A tracking accuracy of 4.5 pixels (approximately 1.5 cm) could be achieved by adding 15-20 manually labeled frames per video. The MAEs for the trained models ranged from 0.21 ± 0.15 for arm dystonia to 0.14 ± 0.10 for leg dystonia (normalized DIS scores). The inter-rater MAEs were 0.21 ± 0.22 and 0.16 ± 0.20, respectively. This proof-of-concept study shows the potential of using stick figures extracted from common videos in a machine learning approach to automatically assess dystonia. Sufficient tracking accuracy can be reached by manually adding labels within 15-20 frames per video. With a relatively small data set, it is possible to train a model that can automatically assess dystonia with a performance comparable to human scoring.
基于视频的临床评分在评估肌张力障碍以及监测运动障碍型脑瘫(CP)的治疗效果方面发挥着重要作用。然而,临床医生进行评估耗时较长,且评分质量取决于经验。本研究的目的是为一种机器学习方法提供概念验证,该方法利用从视频中提取的二维人体简笔画自动评估肌张力障碍评分。将模型性能与人类表现进行比较。共拍摄了34名运动障碍型脑瘫患者(8 - 23岁,均无法行走)在卧位和支撑坐位休息时的187个视频序列。三位评估者根据手臂和腿部肌张力障碍的运动障碍损害量表(DIS)对视频进行评分(标准化分数范围为0 - 1)。使用DeepLabCut提取左手腕、右手腕、肘部、肩部、髋部、膝盖和脚踝的像素坐标,DeepLabCut是一个基于姿态估计算法的开源工具箱。在一个子集中,评估了预训练人体模型以及使用越来越多手动标注帧训练的模型的跟踪准确性。计算了DeepLabCut对身体各点位置的预测与手动标注之间的平均绝对误差(MAE)。随后,根据提取的身体各点坐标计算运动和位置特征。将这些特征输入随机森林回归器以训练一个模型来预测临床评分。将用一位评估者的数据训练的模型通过MAE评估的性能(模型 - 评估者)与评估者间的准确性进行比较。每个视频添加15 - 20个手动标注帧可实现4.5像素(约1.5厘米)的跟踪准确性。训练模型的MAE范围从手臂肌张力障碍的0.21±0.15到腿部肌张力障碍的0.14±0.10(DIS标准化分数)。评估者间的MAE分别为0.21±0.22和0.16±0.20。这项概念验证研究表明了在机器学习方法中使用从普通视频中提取的人体简笔画自动评估肌张力障碍的潜力。通过每个视频手动添加15 - 20帧标注可达到足够的跟踪准确性。利用相对较小的数据集,有可能训练出一个能够自动评估肌张力障碍的模型,其性能可与人类评分相媲美。