File Download

There are no files associated with this item.

Supplementary

Conference Paper: Risk Bounds of Multi-Pass SGD for Least Squares in the Interpolation Regime

TitleRisk Bounds of Multi-Pass SGD for Least Squares in the Interpolation Regime
Authors
Issue Date12-Dec-2022
Abstract

Stochastic gradient descent (SGD) has achieved great success due to its superior performance in both optimization and generalization. Most of existing generalization analyses are made for single-pass SGD, which is a less practical variant compared to the commonly-used multi-pass SGD. Besides, theoretical analyses for multi-pass SGD often concern a worst-case instance in a class of problems, which may be pessimistic to explain the superior generalization ability for some particular problem instance. The goal of this paper is to sharply characterize the generalization of multi-pass SGD, by developing an instance-dependent excess risk bound for least squares in the interpolation regime, which is expressed as a function of the iteration number, stepsize, and data covariance. We show that the excess risk of SGD can be exactly decomposed into the excess risk of GD and a positive fluctuation error, suggesting that SGD always performs worse, instance-wisely, than GD, in generalization. On the other hand, we show that although SGD needs more iterations than GD to achieve the same level of excess risk, it saves the number of stochastic gradient evaluations, and therefore is preferable in terms of computational time.


Persistent Identifierhttp://hdl.handle.net/10722/339354

 

DC FieldValueLanguage
dc.contributor.authorZou, Difan-
dc.contributor.authorWu, Jingfeng-
dc.contributor.authorBraverman, Vladimir-
dc.contributor.authorGu, Quanquan-
dc.contributor.authorKakade, Sham M -
dc.date.accessioned2024-03-11T10:35:56Z-
dc.date.available2024-03-11T10:35:56Z-
dc.date.issued2022-12-12-
dc.identifier.urihttp://hdl.handle.net/10722/339354-
dc.description.abstract<p>Stochastic gradient descent (SGD) has achieved great success due to its superior performance in both optimization and generalization. Most of existing generalization analyses are made for single-pass SGD, which is a less practical variant compared to the commonly-used multi-pass SGD. Besides, theoretical analyses for multi-pass SGD often concern a worst-case instance in a class of problems, which may be pessimistic to explain the superior generalization ability for some particular problem instance. The goal of this paper is to sharply characterize the generalization of multi-pass SGD, by developing an instance-dependent excess risk bound for least squares in the interpolation regime, which is expressed as a function of the iteration number, stepsize, and data covariance. We show that the excess risk of SGD can be exactly decomposed into the excess risk of GD and a positive fluctuation error, suggesting that SGD always performs worse, instance-wisely, than GD, in generalization. On the other hand, we show that although SGD needs more iterations than GD to achieve the same level of excess risk, it saves the number of stochastic gradient evaluations, and therefore is preferable in terms of computational time.</p>-
dc.languageeng-
dc.relation.ispartofAdvances in Neural Information Processing Systems (28/11/2022-09/12/2022, New Orleans)-
dc.titleRisk Bounds of Multi-Pass SGD for Least Squares in the Interpolation Regime-
dc.typeConference_Paper-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats