Department of Neurosurgery and Neurooncology, First Faculty of Medicine, Charles University, Prague, Czech Republic.
Department of Psychiatry, Faculty of Medicine in Pilsen, Charles University, Pilsen, Czech Republic.
J Med Internet Res. 2023 May 31;25:e46924. doi: 10.2196/46924.
Artificial intelligence (AI) has advanced substantially in recent years, transforming many industries and improving the way people live and work. In scientific research, AI can enhance the quality and efficiency of data analysis and publication. However, AI has also opened up the possibility of generating high-quality fraudulent papers that are difficult to detect, raising important questions about the integrity of scientific research and the trustworthiness of published papers.
The aim of this study was to investigate the capabilities of current AI language models in generating high-quality fraudulent medical articles. We hypothesized that modern AI models can create highly convincing fraudulent papers that can easily deceive readers and even experienced researchers.
This proof-of-concept study used ChatGPT (Chat Generative Pre-trained Transformer) powered by the GPT-3 (Generative Pre-trained Transformer 3) language model to generate a fraudulent scientific article related to neurosurgery. GPT-3 is a large language model developed by OpenAI that uses deep learning algorithms to generate human-like text in response to prompts given by users. The model was trained on a massive corpus of text from the internet and is capable of generating high-quality text in a variety of languages and on various topics. The authors posed questions and prompts to the model and refined them iteratively as the model generated the responses. The goal was to create a completely fabricated article including the abstract, introduction, material and methods, discussion, references, charts, etc. Once the article was generated, it was reviewed for accuracy and coherence by experts in the fields of neurosurgery, psychiatry, and statistics and compared to existing similar articles.
The study found that the AI language model can create a highly convincing fraudulent article that resembled a genuine scientific paper in terms of word usage, sentence structure, and overall composition. The AI-generated article included standard sections such as introduction, material and methods, results, and discussion, as well a data sheet. It consisted of 1992 words and 17 citations, and the whole process of article creation took approximately 1 hour without any special training of the human user. However, there were some concerns and specific mistakes identified in the generated article, specifically in the references.
The study demonstrates the potential of current AI language models to generate completely fabricated scientific articles. Although the papers look sophisticated and seemingly flawless, expert readers may identify semantic inaccuracies and errors upon closer inspection. We highlight the need for increased vigilance and better detection methods to combat the potential misuse of AI in scientific research. At the same time, it is important to recognize the potential benefits of using AI language models in genuine scientific writing and research, such as manuscript preparation and language editing.
人工智能(AI)近年来取得了实质性进展,改变了许多行业,并提高了人们的生活和工作方式。在科学研究中,AI 可以提高数据分析和发表的质量和效率。然而,AI 也开辟了生成高质量欺诈性论文的可能性,这些论文很难被发现,这就提出了关于科学研究诚信和已发表论文可信度的重要问题。
本研究旨在调查当前 AI 语言模型在生成高质量医疗欺诈性文章方面的能力。我们假设现代 AI 模型可以生成极具说服力的欺诈性论文,这些论文很容易欺骗读者,甚至是有经验的研究人员。
这是一项概念验证研究,使用 ChatGPT(基于 GPT-3 的聊天生成预训练转换器)生成与神经外科相关的虚假科学文章。GPT-3 是由 OpenAI 开发的大型语言模型,它使用深度学习算法根据用户提供的提示生成类似人类的文本。该模型经过来自互联网的大量文本语料库的训练,能够用各种语言和各种主题生成高质量的文本。作者向模型提出问题和提示,并在模型生成回复时迭代地对其进行细化。目标是创建一篇完全虚构的文章,包括摘要、引言、材料和方法、讨论、参考文献、图表等。文章生成后,由神经外科、精神病学和统计学领域的专家对其准确性和连贯性进行审查,并与现有类似文章进行比较。
研究发现,AI 语言模型可以创建一篇极具说服力的欺诈性文章,在词汇使用、句子结构和整体组成方面与真正的科学论文相似。AI 生成的文章包括引言、材料和方法、结果和讨论等标准部分,以及数据表。它由 1992 个单词和 17 个引文组成,整个文章创建过程大约需要 1 小时,而无需对人类用户进行任何特殊培训。然而,在生成的文章中发现了一些关注和特定的错误,特别是在参考文献中。
该研究表明,当前的 AI 语言模型有潜力生成完全虚构的科学文章。虽然这些论文看起来很复杂,似乎无懈可击,但有经验的读者在仔细检查时可能会发现语义上的不准确和错误。我们强调需要提高警惕并开发更好的检测方法来打击 AI 在科学研究中可能的滥用。同时,认识到在真正的科学写作和研究中使用 AI 语言模型的潜在好处也很重要,例如手稿准备和语言编辑。