2000 character limit reached
Spectral Barron space for deep neural network approximation
Published 2 Sep 2023 in math.NA and cs.NA | (2309.00788v2)
Abstract: We prove the sharp embedding between the spectral Barron space and the Besov space with embedding constants independent of the input dimension. Given the spectral Barron space as the target function space, we prove a dimension-free convergence result that if the neural network contains $L$ hidden layers with $N$ units per layer, then the upper and lower bounds of the $L2$-approximation error are $\mathcal{O}(N{-sL})$ with $0 < sL\le 1/2$, where $s\ge 0$ is the smoothness index of the spectral Barron space.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.