Click here to flash read.
arXiv:2304.10552v2 Announce Type: replace
Abstract: In this paper, we prove that in the overparametrized regime, deep neural network provide universal approximations and can interpolate any data set, as long as the activation function is locally in $L^1(\RR)$ and not an affine function.
Additionally, if the activation function is smooth and such an interpolation networks exists, then the set of parameters which interpolate forms a manifold. Furthermore, we give a characterization of the Hessian of the loss function evaluated at the interpolation points.
In the last section, we provide a practical probabilistic method of finding such a point under general conditions on the activation function.
Click here to read this post out
ID: 822588; Unique Viewers: 0
Unique Voters: 0
Total Votes: 0
Votes:
Latest Change: April 26, 2024, 7:31 a.m.
Changes:
Dictionaries:
Words:
Spaces:
Views: 9
CC:
No creative common's license
No creative common's license
Comments: