Nonlinear functional regression by functional deep neural network with kernel embedding
Abstract: Recently, deep learning has been widely applied in functional data analysis (FDA) with notable empirical success. However, the infinite dimensionality of functional data necessitates an effective dimension reduction approach for functional learning tasks, particularly in nonlinear functional regression. In this paper, we introduce a functional deep neural network with an adaptive and discretization-invariant dimension reduction method. Our functional network architecture consists of three parts: first, a kernel embedding step that features an integral transformation with an adaptive smooth kernel; next, a projection step that utilizes eigenfunction bases based on a projection Mercer kernel for the dimension reduction; and finally, a deep ReLU neural network is employed for the prediction. Explicit rates of approximating nonlinear smooth functionals across various input function spaces by our proposed functional network are derived. Additionally, we conduct a generalization analysis for the empirical risk minimization (ERM) algorithm applied to our functional net, by employing a novel two-stage oracle inequality and the established functional approximation results. Ultimately, we conduct numerical experiments on both simulated and real datasets to demonstrate the effectiveness and benefits of our functional net.
- Sobolev Spaces. Elsevier, 2003.
- Neural operator: Graph kernel network for partial differential equations. In ICLR 2020 Workshop on Integration of Deep Neural Models and Differential Equations, 2020.
- Francis Bach. On the equivalence between kernel quadrature rules and random feature expansions. The Journal of Machine Learning Research, 18(1):714–751, 2017.
- Nearly-tight VC-dimension and pseudodimension bounds for piecewise linear neural networks. The Journal of Machine Learning Research, 20(1):2285–2301, 2019.
- Reproducing Kernel Hilbert Spaces in Probability and Statistics. Springer Science & Business Media, 2011.
- Principal components analysis of sampled functions. Psychometrika, 51(2):285–311, 1986.
- Model reduction and neural networks for parametric PDEs. The SMAI Journal of Computational Mathematics, 7:121–157, 2021.
- Spline estimators for the functional linear model. Statistica Sinica, pages 571–591, 2003.
- Universal approximation to nonlinear operators by neural networks with arbitrary activation functions and its application to dynamical systems. IEEE Transactions on Neural Networks, 6(4):911–917, 1995.
- Online gradient descent algorithms for functional data learning. Journal of Complexity, 70:101635, 2022.
- Deep neural networks for rotation-invariance approximation and learning. Analysis and Applications, 17(05):737–772, 2019.
- Moo K Chung. Statistical and Computational Methods in Brain Image Analysis. CRC press, 2013.
- On the mathematical foundations of learning. Bulletin of the American Mathematical Society, 39(1):1–49, 2002.
- Learning Theory: An Approximation Theory Viewpoint, volume 24. Cambridge University Press, 2007.
- Learning to synthesize: robust phase retrieval at low photon counts. Light: Science & Applications, 9(1):36, 2020.
- Constructive Approximation, volume 303. Springer Science & Business Media, 1993.
- Optimal regression rates for SVMs using Gaussian kernels. Electronic Journal of Statistics, 7:1–42, 2013.
- Gregory E Fasshauer. Positive definite kernels: past, present and future. Dolomites Research Notes on Approximation, 4:21–63, 2011.
- Nonparametric Functional Data Analysis: Theory and Practice, volume 76. New York: Springer, 2006.
- A Distribution-Free Theory of Nonparametric Regression, volume 1. Springer, 2002.
- David Haussler. Decision theoretic generalizations of the PAC model for neural net and other learning applications. Information and Computation, 100(1):78–150, 1992.
- Solving parametric pde problems with artificial neural networks. European Journal of Applied Mathematics, 32(3):421–435, 2021.
- On universal approximation and error bounds for fourier neural operators. The Journal of Machine Learning Research, 22(1):13237–13312, 2021.
- Error estimates for DeepONets: A deep learning framework in infinite dimensions. Transactions of Mathematics and Its Applications, 6(1):tnac001, 2022.
- Fourier neural operator for parametric partial differential equations. In International Conference on Learning Representations, 2021.
- Multipole graph neural operator for parametric partial differential equations. Advances in Neural Information Processing Systems, 33:6755–6766, 2020.
- Learning nonlinear operators via DeepONet based on the universal approximation theorem of operators. Nature Machine Intelligence, 3(3):218–229, 2021.
- Theory of deep convolutional neural networks III: Approximating radial functions. Neural Networks, 144:778–790, 2021.
- Approximating functions with multi-features by deep convolutional neural networks. Analysis and Applications, pages 1–33, 2022.
- Neural networks for functional approximation and system identification. Neural Computation, 9(1):143–159, 1997.
- New error bounds for deep ReLU networks using sparse grids. SIAM Journal on Mathematics of Data Science, 1(1):78–92, 2019.
- Integral autoencoder network for discretization-invariant learning. The Journal of Machine Learning Research, 23(286):1–45, 2022.
- Allan Pinkus. N-widths in Approximation Theory, volume 7. Springer Science & Business Media, 2012.
- Numerical Recipes in C. Cambridge University Press, 1992.
- Evaluation and development of deep neural networks for image super-resolution in optical microscopy. Nature Methods, 18(2):194–202, 2021.
- Image inpainting based on deep learning: A review. Displays, 69:102028, 2021.
- scikit-fda: a Python package for functional data analysis. arXiv preprint arXiv:2211.02566, 2022.
- Functional Data Analysis. New York: Springer, 2005.
- Carl Edward Rasmussen and Christopher K. I. Williams. Gaussian Processes for Machine Learning. MIT press, 2006.
- Nonparametric mixed effects models for unequally sampled noisy curves. Biometrics, 57(1):253–259, 2001.
- Functional multi-layer perceptron: a non-linear tool for functional data analysis. Neural Networks, 18(1):45–60, 2005.
- Functional data analysis with multi layer perceptrons. In Proceedings of the 2002 International Joint Conference on Neural Networks. IJCNN’02 (Cat. No. 02CH37290), volume 3, pages 2843–2848. IEEE, 2002.
- Representation of functional data in neural networks. Neurocomputing, 64:183–210, 2005.
- Johannes Schmidt-Hieber. Nonparametric regression using deep neural networks with ReLU activation function. The Annals of Statistics, 48(4):1875–1897, 2020.
- Bernard W Silverman. Smoothed functional principal components analysis by choice of norm. The Annals of Statistics, 24(1):1–24, 1996.
- A Hilbert space embedding for distributions. In Algorithmic Learning Theory, pages 13–31. Springer, 2007.
- Approximation of nonlinear functionals using deep ReLU networks. arXiv preprint arXiv:2304.04443, 2023.
- Approximation of smooth functionals using deep ReLU networks. Neural Networks, 2022. Minor revision.
- Support Vector Machines. Springer Science & Business Media, 2008.
- Taiji Suzuki. Adaptivity of deep ReLU network for learning in Besov and mixed smooth Besov spaces: optimal rate and curse of dimensionality. In International Conference on Learning Representations, 2019.
- Matus Telgarsky. Benefits of depth in neural networks. In Conference on Learning Theory, pages 1517–1539. PMLR, 2016.
- Deep learning on image denoising: An overview. Neural Networks, 131:251–275, 2020.
- Kernel Smoothing. CRC press, 1994.
- Holger Wendland. Scattered Data Approximation, volume 17. Cambridge university press, 2004.
- Functional data analysis for sparse longitudinal data. Journal of the American Statistical Association, 100(470):577–590, 2005.
- Deep learning for functional data analysis with adaptive basis layers. In International Conference on Machine Learning, pages 11898–11908. PMLR, 2021.
- Dmitry Yarotsky. Error bounds for approximations with deep ReLU networks. Neural Networks, 94:103–114, 2017.
- Dmitry Yarotsky. Optimal approximation of continuous functions by very deep ReLU networks. In Conference on Learning Theory, pages 639–649. PMLR, 2018.
- Ding-Xuan Zhou. Universality of deep convolutional neural networks. Applied and Computational Harmonic Analysis, 48(2):787–794, 2020.
- Functional linear regression for discretely observed data: from ideal to reality. Biometrika, 2022.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.