Havers Tim, Masur Lukas, Isenmann Eduard, Geisler Stephan, Zinner Christoph, Sperlich Billy, Düking Peter
Department of Fitness and Health, IST University of Applied Sciences, Düsseldorf, Germany.
Faculty of Sport and Health Sciences, Technical University of Munich, Munich, Germany.
Biol Sport. 2025 Apr;42(2):289-329. doi: 10.5114/biolsport.2025.145911. Epub 2024 Dec 18.
Large Language Models (LLMs) are increasingly utilized in various domains, including the generation of training plans. However, the reproducibility and quality of training plans produced by different LLMs have not been studied extensively. This study aims to: i) investigate and compare the quality of muscle hypertrophy-related resistance training (RT) plans generated by Google Gemini (GG) and GPT-4, and ii) the reproducibility of the RT plans when the same prompts are provided multiple times concomitantly. Two distinct prompts were used, one providing little information about the training plan requirements and the other providing detailed information. These prompts were input into GG and GPT-4 by two different individuals, resulting in the generation of eight RT plans. These plans were evaluated by 12 coaching experts using a 5-point Likert scale, based on quality criteria derived from the literature. The results indicated a high degree of reproducibility, as indicated by coaching expert evaluation, when the same distinct prompts were provided multiple times to the LLMs of interest, with 27 out of 28 items showing no differences (p > 0.05). Overall, GPT-4 was rated higher on several aspects of RT quality criteria (p = 0.000-0.043). Additionally, compared to little information, higher information density within the prompts resulted in higher rated RT quality (p = 0.000-0.037). Our findings show that RT plans can be generated reproducibly with the same quality when using the same prompts. Furthermore, quality improves with more detailed input, and GPT-4 outperformed GG in generating higherquality plans. These results suggest that detailed information input is crucial for LLM performance.
大语言模型(LLMs)越来越多地应用于各个领域,包括训练计划的生成。然而,不同大语言模型生成的训练计划的可重复性和质量尚未得到广泛研究。本研究旨在:i)调查并比较由谷歌Gemini(GG)和GPT-4生成的与肌肉肥大相关的阻力训练(RT)计划的质量,以及ii)当多次同时提供相同提示时RT计划的可重复性。使用了两个不同的提示,一个提供关于训练计划要求的信息很少,另一个提供详细信息。这两个提示由两个不同的人输入到GG和GPT-4中,从而生成了八个RT计划。这些计划由12名教练专家根据从文献中得出的质量标准,使用5点李克特量表进行评估。结果表明,当向相关大语言模型多次提供相同的不同提示时,教练专家评估显示出高度的可重复性,28个项目中有27个没有差异(p>0.05)。总体而言,GPT-4在RT质量标准的几个方面得分更高(p=0.000 - 0.043)。此外,与信息少相比,提示中的信息密度更高会导致RT质量评分更高(p=0.000 - 0.037)。我们的研究结果表明,使用相同提示可以可重复地生成具有相同质量的RT计划。此外,随着输入更详细,质量会提高,并且GPT-4在生成更高质量计划方面优于GG。这些结果表明,详细的信息输入对大语言模型的性能至关重要。
J Funct Morphol Kinesiol. 2023-12-29
Front Psychiatry. 2024-1-4
Med Sci Sports Exerc. 2024-3-1
Int J Sports Physiol Perform. 2023-10-1
Nat Med. 2023-8