title | abstract | layout | series | publisher | issn | id | month | tex_title | firstpage | lastpage | page | order | cycles | bibtex_author | author | date | address | container-title | volume | genre | issued | extras | ||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
**Paper retracted by author request (see pdf for retraction notice from the authors)** Nonparametric Regression with Shallow Overparameterized Neural Networks Trained by GD with Early Stopping |
We explore the ability of overparameterized shallow neural networks to learn Lipschitz regression functions with and without label noise when trained by Gradient Descent (GD). To avoid the problem that in the presence of noisy labels, neural networks trained to nearly zero training error are inconsistent on this class, we propose an early stopping rule that allows us to show optimal rates. This provides an alternative to the result of Hu et al. (2021) who studied the performance of |
inproceedings |
Proceedings of Machine Learning Research |
PMLR |
2640-3498 |
kuzborskij21a |
0 |
Nonparametric Regression with Shallow Overparameterized Neural Networks Trained by GD with Early Stopping |
2853 |
2890 |
2853-2890 |
2853 |
false |
Kuzborskij, Ilja and Szepesvari, Csaba |
|
2021-07-21 |
Proceedings of Thirty Fourth Conference on Learning Theory |
134 |
inproceedings |
|