School of Mathematical Science, Beihang University, Beijing, China; School of Data Science, City University of Hong Kong, Kowloon, Hong Kong.
Laboratory for AI-Powered Financial Technologies, Hong Kong Science Park, Shatin, New Territories, Hong Kong.
Neural Netw. 2023 Sep;166:424-436. doi: 10.1016/j.neunet.2023.07.012. Epub 2023 Jul 18.
In recent years, deep neural networks have been employed to approximate nonlinear continuous functionals F defined on L([-1,1]) for 1≤p≤∞. However, the existing theoretical analysis in the literature either is unsatisfactory due to the poor approximation results, or does not apply to the rectified linear unit (ReLU) activation function. This paper aims to investigate the approximation power of functional deep ReLU networks in two settings: F is continuous with restrictions on the modulus of continuity, and F has higher order Fréchet derivatives. A novel functional network structure is proposed to extract features of higher order smoothness harbored by the target functional F. Quantitative rates of approximation in terms of the depth, width and total number of weights of neural networks are derived for both settings. We give logarithmic rates when measuring the approximation error on the unit ball of a Hölder space. In addition, we establish nearly polynomial rates (i.e., rates of the form exp-a(logM) with a>0,0<b<1) when measuring the approximation error on a space of analytic functions.
近年来,深度神经网络已被用于逼近定义在 L([-1,1]) 上的 1≤p≤∞ 的非线性连续泛函 F。然而,文献中的现有理论分析要么由于逼近效果不佳而不令人满意,要么不适用于修正线性单元 (ReLU) 激活函数。本文旨在研究函数深度 ReLU 网络在两种情况下的逼近能力:F 是连续的,且在连续性模上有约束,F 具有高阶 Fréchet 导数。提出了一种新的函数网络结构,以提取目标函数 F 所具有的更高阶平滑性特征。针对这两种情况,推导出了网络深度、宽度和总权重数量的定量逼近率。当在 Hölder 空间的单位球上度量逼近误差时,我们给出了对数率。此外,当在解析函数空间上度量逼近误差时,我们建立了几乎多项式率(即形如 exp-a(logM) 的率,其中 a>0,0<b<1)。