File Download

There are no files associated with this item.

Supplementary

Conference Paper: Learning High-Dimensional Single-Neuron ReLU Networks with Finite Samples

TitleLearning High-Dimensional Single-Neuron ReLU Networks with Finite Samples
Authors
Issue Date23-Jul-2022
Abstract

This paper considers the problem of learning a single ReLU neuron with squared loss (a.k.a., ReLU regression) in the overparameterized regime, where the input dimension can exceed the number of samples. We analyze a Perceptron-type algorithm called GLM-tron (Kakade et al., 2011) and provide its dimension-free risk upper bounds for high-dimensional ReLU regression in both well-specified and misspecified settings. Our risk bounds recover several existing results as special cases. Moreover, in the well-specified setting, we provide an instance-wise matching risk lower bound for GLM-tron. Our upper and lower risk bounds provide a sharp characterization of the high-dimensional ReLU regression problems that can be learned via GLM-tron. On the other hand, we provide some negative results for stochastic gradient descent (SGD) for ReLU regression with symmetric Bernoulli data: if the model is wellspecified, the excess risk of SGD is provably no better than that of GLM-tron ignoring constant factors, for each problem instance; and in the noiseless case, GLM-tron can achieve a small risk while SGD unavoidably suffers from a constant risk in expectation. These results together suggest that GLM-tron might be preferable to SGD for high-dimensional ReLU regression.


Persistent Identifierhttp://hdl.handle.net/10722/339355

 

DC FieldValueLanguage
dc.contributor.authorWu, Jingfeng-
dc.contributor.authorZou, Difan-
dc.contributor.authorChen, Zixiang-
dc.contributor.authorBraverman, Vladimir-
dc.contributor.authorGu, Quanquan-
dc.contributor.authorKakade, Sham M -
dc.date.accessioned2024-03-11T10:35:57Z-
dc.date.available2024-03-11T10:35:57Z-
dc.date.issued2022-07-23-
dc.identifier.urihttp://hdl.handle.net/10722/339355-
dc.description.abstract<p>This paper considers the problem of learning a single ReLU neuron with squared loss (a.k.a., ReLU regression) in the overparameterized regime, where the input dimension can exceed the number of samples. We analyze a Perceptron-type algorithm called GLM-tron (Kakade et al., 2011) and provide its dimension-free risk upper bounds for high-dimensional ReLU regression in both well-specified and misspecified settings. Our risk bounds recover several existing results as special cases. Moreover, in the well-specified setting, we provide an instance-wise matching risk lower bound for GLM-tron. Our upper and lower risk bounds provide a sharp characterization of the high-dimensional ReLU regression problems that can be learned via GLM-tron. On the other hand, we provide some negative results for stochastic gradient descent (SGD) for ReLU regression with symmetric Bernoulli data: if the model is wellspecified, the excess risk of SGD is provably no better than that of GLM-tron ignoring constant factors, for each problem instance; and in the noiseless case, GLM-tron can achieve a small risk while SGD unavoidably suffers from a constant risk in expectation. These results together suggest that GLM-tron might be preferable to SGD for high-dimensional ReLU regression.</p>-
dc.languageeng-
dc.relation.ispartofInternational Conference on Machine Learning (17/07/2022-23/07/2022, Baltimore)-
dc.titleLearning High-Dimensional Single-Neuron ReLU Networks with Finite Samples-
dc.typeConference_Paper-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats