File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Fine-Grained analysis of stability and generalization for stochastic gradientdescent

TitleFine-Grained analysis of stability and generalization for stochastic gradientdescent
Authors
Issue Date2020
Citation
37th International Conference on Machine Learning, ICML 2020, 2020, v. PartF168147-8, p. 5765-5775 How to Cite?
AbstractRecently there are a considerable amount of work devoted to the study of the algorithmic stability and generalization for stochastic gradient descent (SGD). However, the existing stability analysis requires to impose restrictive assumptions on the boundedness of gradients, smoothness and con_vexity of loss functions. In this paper, we provide a fine-grained analysis of stability and general_ization for SGD by substantially relaxing these assumptions. Firstly, we establish stability and generalization for SGD by removing the existing bounded gradient assumptions. The key idea is the introduction of a new stability measure called on-average model stability, for which we develop novel bounds controlled by the risks of SGD iter_ates. This yields generalization bounds depend_ing on the behavior of the best model, and leads to the first-ever-known fast bounds in the low_noise setting using stability approach. Secondly, the smoothness assumption is relaxed by con_sidering loss functions with Holder continuous (sub)gradients for which we show that optimal bounds are still achieved by balancing computa_tion and stability. To our best knowledge, this gives the first-ever-known stability and generaliza_tion bounds for SGD with non-smooth loss func_tions (e.g., hinge loss). Finally, we study learning problems with (strongly) convex objectives but non-convex loss functions.
Persistent Identifierhttp://hdl.handle.net/10722/329705

 

DC FieldValueLanguage
dc.contributor.authorLei, Yunwen-
dc.contributor.authorYing, Yiming-
dc.date.accessioned2023-08-09T03:34:44Z-
dc.date.available2023-08-09T03:34:44Z-
dc.date.issued2020-
dc.identifier.citation37th International Conference on Machine Learning, ICML 2020, 2020, v. PartF168147-8, p. 5765-5775-
dc.identifier.urihttp://hdl.handle.net/10722/329705-
dc.description.abstractRecently there are a considerable amount of work devoted to the study of the algorithmic stability and generalization for stochastic gradient descent (SGD). However, the existing stability analysis requires to impose restrictive assumptions on the boundedness of gradients, smoothness and con_vexity of loss functions. In this paper, we provide a fine-grained analysis of stability and general_ization for SGD by substantially relaxing these assumptions. Firstly, we establish stability and generalization for SGD by removing the existing bounded gradient assumptions. The key idea is the introduction of a new stability measure called on-average model stability, for which we develop novel bounds controlled by the risks of SGD iter_ates. This yields generalization bounds depend_ing on the behavior of the best model, and leads to the first-ever-known fast bounds in the low_noise setting using stability approach. Secondly, the smoothness assumption is relaxed by con_sidering loss functions with Holder continuous (sub)gradients for which we show that optimal bounds are still achieved by balancing computa_tion and stability. To our best knowledge, this gives the first-ever-known stability and generaliza_tion bounds for SGD with non-smooth loss func_tions (e.g., hinge loss). Finally, we study learning problems with (strongly) convex objectives but non-convex loss functions.-
dc.languageeng-
dc.relation.ispartof37th International Conference on Machine Learning, ICML 2020-
dc.titleFine-Grained analysis of stability and generalization for stochastic gradientdescent-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.scopuseid_2-s2.0-85105548031-
dc.identifier.volumePartF168147-8-
dc.identifier.spage5765-
dc.identifier.epage5775-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats