Efficient Approximation to Analytic and $L^p$ functions by Height-Augmented ReLU Networks
About
This work addresses two fundamental limitations in neural network approximation theory. We demonstrate that a three-dimensional network architecture enables a significantly more efficient representation of sawtooth functions, which serves as the cornerstone in the approximation of analytic and $L^p$ functions. First, we establish substantially improved exponential approximation rates for several important classes of analytic functions and offer a parameter-efficient network design. Second, for the first time, we derive a quantitative and non-asymptotic approximation of high orders for general $L^p$ functions. Our techniques advance the theoretical understanding of the neural network approximation in fundamental function spaces and offer a theoretically grounded pathway for designing more parameter-efficient networks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Function Approximation | Analytic functions on [0, 1 - δ]^d | Approximation Error1 | 2 | |
| Function Approximation | Analytic functions on L²(ℝ^d, γ_d), holomorphic in a strip | Error Bound1 | 2 | |
| Function Approximation | Polynomial functions on [0, 1] | Error2 | 1 | |
| Function Approximation | Lᵖ functions on [−1, 1]^d | Error1 | 1 | |
| Function Approximation | Analytic functions on [0, 1]^d, holomorphic in an ellipse | -- | 1 |