Kakarmath Sujay, Esteva Andre, Arnaout Rima, Harvey Hugh, Kumar Santosh, Muse Evan, Dong Feng, Wedlund Leia, Kvedar Joseph
MGH & BWH Center for Clinical Data Science, Partners Healthcare, Boston, MA USA.
Department of Medical AI, Salesforce Research, Palo Alto, CA USA.
NPJ Digit Med. 2020 Oct 16;3:134. doi: 10.1038/s41746-020-00336-w. eCollection 2020.
Since its inception in 2017, has attracted a disproportionate number of manuscripts reporting on uses of artificial intelligence. This field has matured rapidly in the past several years. There was initial fascination with the algorithms themselves (machine learning, deep learning, convoluted neural networks) and the use of these algorithms to make predictions that often surpassed prevailing benchmarks. As the discipline has matured, individuals have called attention to aberrancies in the output of these algorithms. In particular, criticisms have been widely circulated that algorithmically developed models may have limited generalizability due to overfitting to the training data and may systematically perpetuate various forms of biases inherent in the training data, including race, gender, age, and health state or fitness level (Challen et al. BMJ Qual. Saf. 28:231-237, 2019; O'neil. Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy, Broadway Book, 2016). Given our interest in publishing the highest quality papers and the growing volume of submissions using AI algorithms, we offer a list of criteria that authors should consider before submitting papers to .
自2017年创立以来,已吸引了数量极不相称的关于人工智能应用的稿件。在过去几年中,这个领域迅速成熟。最初人们着迷于算法本身(机器学习、深度学习、卷积神经网络)以及使用这些算法进行预测,这些预测常常超过了现有的基准。随着该学科的成熟,人们开始关注这些算法输出中的异常情况。特别是,有广泛流传的批评称,通过算法开发的模型可能由于过度拟合训练数据而具有有限的通用性,并且可能会系统性地延续训练数据中固有的各种形式的偏差,包括种族、性别、年龄以及健康状况或 fitness 水平(Challen等人,《英国医学杂志·质量与安全》28:231 - 237,2019;奥尼尔,《数学毁灭武器:大数据如何加剧不平等并威胁民主》,百老汇图书,2016)。鉴于我们对发表最高质量论文的兴趣以及使用人工智能算法的投稿数量不断增加,我们提供了一份作者在向投稿前应考虑的标准清单。