The School of Computing, University of Kent, United Kingdom.
Manada Technology LLC, United States.
Neurosci Biobehav Rev. 2020 Dec;119:456-467. doi: 10.1016/j.neubiorev.2020.09.036. Epub 2020 Oct 6.
Machine learning has enhanced the abilities of neuroscientists to interpret information collected through EEG, fMRI, and MEG data. With these powerful techniques comes the danger of overfitting of hyperparameters which can render results invalid. We refer to this problem as 'overhyping' and show that it is pernicious despite commonly used precautions. Overhyping occurs when analysis decisions are made after observing analysis outcomes and can produce results that are partially or even completely spurious. It is commonly assumed that cross-validation is an effective protection against overfitting or overhyping, but this is not actually true. In this article, we show that spurious results can be obtained on random data by modifying hyperparameters in seemingly innocuous ways, despite the use of cross-validation. We recommend a number of techniques for limiting overhyping, such as lock boxes, blind analyses, pre-registrations, and nested cross-validation. These techniques, are common in other fields that use machine learning, including computer science and physics. Adopting similar safeguards is critical for ensuring the robustness of machine-learning techniques in the neurosciences.
机器学习增强了神经科学家解读通过 EEG、fMRI 和 MEG 数据收集的信息的能力。这些强大的技术带来了超参数过度拟合的危险,这可能会导致结果无效。我们将这个问题称为“过度炒作”,并表明尽管通常采取了预防措施,它仍然是有害的。过度炒作发生在观察分析结果后做出分析决策时,可能会产生部分甚至完全虚假的结果。人们通常认为交叉验证是防止过度拟合或过度炒作的有效方法,但实际上并非如此。在本文中,我们表明,尽管使用了交叉验证,但通过以看似无害的方式修改超参数,仍然可以在随机数据上获得虚假结果。我们建议采用一些技术来限制过度炒作,例如锁箱、盲分析、预注册和嵌套交叉验证。这些技术在包括计算机科学和物理学在内的其他使用机器学习的领域中很常见。在神经科学中采用类似的保护措施对于确保机器学习技术的稳健性至关重要。