Suppr超能文献

基于视频记录步态的两步深度学习足跟关键点识别

Two-step deep-learning identification of heel keypoints from video-recorded gait.

作者信息

Halvorsen Kjartan, Peng Wei, Olsson Fredrik, Åberg Anna Cristina

机构信息

School of Health and Welfare, Dalarna University, Falun, Sweden.

Department of Public Health and Caring Sciences, Uppsala University, Uppsala, Sweden.

出版信息

Med Biol Eng Comput. 2025 Jan;63(1):229-237. doi: 10.1007/s11517-024-03189-7. Epub 2024 Sep 18.

Abstract

Accurate and fast extraction of step parameters from video recordings of gait allows for richer information to be obtained from clinical tests such as Timed Up and Go. Current deep-learning methods are promising, but lack in accuracy for many clinical use cases. Extracting step parameters will often depend on extracted landmarks (keypoints) on the feet. We hypothesize that such keypoints can be determined with an accuracy relevant for clinical practice from video recordings by combining an existing general-purpose pose estimation method (OpenPose) with custom convolutional neural networks (convnets) specifically trained to identify keypoints on the heel. The combined method finds keypoints on the posterior and lateral aspects of the heel of the foot in side-view and frontal-view images from which step length and step width can be determined for calibrated cameras. Six different candidate convnets were evaluated, combining three different standard architectures as networks for feature extraction (backbone), and with two different networks for predicting keypoints on the heel (head networks). Using transfer learning, the backbone networks were pre-trained on the ImageNet dataset, and the combined networks (backbone + head) were fine-tuned on data from 184 trials of older, unimpaired adults. The data was recorded at three different locations and consisted of 193 k side-view images and 110 k frontal-view images. We evaluated the six different models using the absolute distance on the floor between predicted keypoints and manually labelled keypoints. For the best-performing convnet, the median error was 0.55 cm and the 75% quartile was below 1.26 cm using data from the side-view camera. The predictions are overall accurate, but show some outliers. The results indicate potential for future clinical use by automating a key step in marker-less gait parameter extraction.

摘要

从步态视频记录中准确快速地提取步幅参数,能够从诸如定时起立行走测试等临床测试中获取更丰富的信息。当前的深度学习方法很有前景,但在许多临床应用案例中准确性不足。提取步幅参数通常依赖于足部提取的地标(关键点)。我们假设,通过将现有的通用姿态估计方法(OpenPose)与专门训练用于识别足跟关键点的定制卷积神经网络(卷积网络)相结合,可以从视频记录中以与临床实践相关的精度确定此类关键点。该组合方法可在侧视图和正视图图像中找到足跟后部和外侧的关键点,据此可以为校准相机确定步长和步宽。对六种不同的候选卷积网络进行了评估,将三种不同的标准架构作为特征提取网络(主干网络),并使用两种不同的网络来预测足跟上的关键点(头部网络)。利用迁移学习,主干网络在ImageNet数据集上进行预训练,组合网络(主干 + 头部)在184例老年健康成年人的试验数据上进行微调。数据在三个不同地点记录,包括19.3万张侧视图图像和11万张正视图图像。我们使用预测关键点与手动标记关键点在地面上的绝对距离来评估这六种不同的模型。对于性能最佳的卷积网络,使用侧视图相机的数据时,中位数误差为0.55厘米,75%四分位数低于1.26厘米。预测总体准确,但存在一些异常值。结果表明,通过自动化无标记步态参数提取中的关键步骤,未来具有临床应用潜力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8a7c/11695559/71efcec7623e/11517_2024_3189_Fig1_HTML.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验