IEEE Trans Neural Netw Learn Syst. 2018 Feb;29(2):500-508. doi: 10.1109/TNNLS.2017.2651985. Epub 2017 Jan 25.
Approximation of high-dimensional functions is a challenge for neural networks due to the curse of dimensionality. Often the data for which the approximated function is defined resides on a low-dimensional manifold and in principle the approximation of the function over this manifold should improve the approximation performance. It has been show that projecting the data manifold into a lower dimensional space, followed by the neural network approximation of the function over this space, provides a more precise approximation of the function than the approximation of the function with neural networks in the original data space. However, if the data volume is very large, the projection into the low-dimensional space has to be based on a limited sample of the data. Here, we investigate the nature of the approximation error of neural networks trained over the projection space. We show that such neural networks should have better approximation performance than neural networks trained on high-dimensional data even if the projection is based on a relatively sparse sample of the data manifold. We also find that it is preferable to use a uniformly distributed sparse sample of the data for the purpose of the generation of the low-dimensional projection. We illustrate these results considering the practical neural network approximation of a set of functions defined on high-dimensional data including real world data as well.
高维函数的逼近是神经网络面临的一个挑战,这是由于维度的诅咒。通常,定义要逼近的函数的数据位于低维流形上,并且原则上,在该流形上逼近函数应该可以提高逼近性能。已经表明,将数据流形投影到较低维空间中,然后在该空间中通过神经网络逼近函数,可以比在原始数据空间中使用神经网络逼近函数提供更精确的逼近。但是,如果数据量非常大,则必须基于数据的有限样本将其投影到低维空间中。在这里,我们研究了在投影空间中训练的神经网络的逼近误差的性质。我们表明,即使投影是基于数据流形的相对稀疏的样本,经过训练的神经网络也应该具有比在高维数据上训练的神经网络更好的逼近性能。我们还发现,为了生成低维投影,最好使用数据的均匀分布的稀疏样本。我们考虑了一组在高维数据(包括真实世界数据)上定义的函数的实际神经网络逼近,来说明这些结果。