Abstract
This paper derives the optimal rate of approximation for Korobov functions with deep neural networks in the high dimensional hypercube with respect to Lp-norms and H1-norm. Our approximation bounds are non-asymptotic in both the width and depth of the networks. The obtained approximation rates demonstrate a remarkable super-convergence feature, improving the existing convergence rates of neural networks that are continuous function approximators. Finally, using a VC-dimension argument, we show that the established rates are near-optimal.
| Original language | English (US) |
|---|---|
| Article number | 106702 |
| Journal | Neural Networks |
| Volume | 180 |
| DOIs | |
| State | Published - Dec 2024 |
Bibliographical note
Publisher Copyright:© 2024 Elsevier Ltd
Keywords
- Deep neural networks
- Korobov spaces
- Optimal approximation rates
- Sobolev spaces
PubMed: MeSH publication types
- Journal Article