Department of Physics, Bar-Ilan University, 52900, Ramat-Gan, Israel.
Gonda Interdisciplinary Brain Research Center, Bar-Ilan University, 52900, Ramat-Gan, Israel.
Sci Rep. 2020 Nov 12;10(1):19628. doi: 10.1038/s41598-020-76764-1.
Power-law scaling, a central concept in critical phenomena, is found to be useful in deep learning, where optimized test errors on handwritten digit examples converge as a power-law to zero with database size. For rapid decision making with one training epoch, each example is presented only once to the trained network, the power-law exponent increased with the number of hidden layers. For the largest dataset, the obtained test error was estimated to be in the proximity of state-of-the-art algorithms for large epoch numbers. Power-law scaling assists with key challenges found in current artificial intelligence applications and facilitates an a priori dataset size estimation to achieve a desired test accuracy. It establishes a benchmark for measuring training complexity and a quantitative hierarchy of machine learning tasks and algorithms.
幂律缩放,临界现象的一个核心概念,在深度学习中被发现是有用的,在那里手写数字样本的优化测试误差随着数据库大小呈幂律收敛到零。对于一个训练周期的快速决策,每个样本只被呈现给训练过的网络一次,幂律指数随着隐藏层的数量增加。对于最大的数据集,获得的测试误差估计接近大epoch 数量的最新算法。幂律缩放有助于解决当前人工智能应用中发现的关键挑战,并有助于进行先验数据集大小估计以达到所需的测试精度。它为衡量训练复杂性和机器学习任务和算法的定量层次结构建立了一个基准。