File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Variance reduction via accelerated dual averaging for finite-sum optimization

TitleVariance reduction via accelerated dual averaging for finite-sum optimization
Authors
Issue Date2020
Citation
Advances in Neural Information Processing Systems, 2020, v. 2020-December How to Cite?
AbstractIn this paper, we introduce a simplified and unified method for finite-sum convex optimization, named Variance Reduction via Accelerated Dual Averaging (VRADA). In both general convex and strongly convex settings, VRADA can attain an O(n1)-accurate solution in O(n log log n) number of stochastic gradient evaluations which improves the best known result O(n log n), where n is the number of samples. Meanwhile, VRADA matches the lower bound of the general convex setting up to a log log n factor and matches the lower bounds in both regimes n < Θ(K) and n > K of the strongly convex setting, where K denotes the condition number. Besides improving the best known results and matching all the above lower bounds simultaneously, VRADA has more unified and simplified algorithmic implementation and convergence analysis for both the general convex and strongly convex settings. The underlying novel approaches such as the novel initialization strategy in VRADA may be of independent interest. Through experiments on real datasets, we show the good performance of VRADA over existing methods for large-scale machine learning problems.
Persistent Identifierhttp://hdl.handle.net/10722/327773
ISSN
2020 SCImago Journal Rankings: 1.399

 

DC FieldValueLanguage
dc.contributor.authorSong, Chaobing-
dc.contributor.authorJinag, Yong-
dc.contributor.authorMa, Yi-
dc.date.accessioned2023-05-08T02:26:43Z-
dc.date.available2023-05-08T02:26:43Z-
dc.date.issued2020-
dc.identifier.citationAdvances in Neural Information Processing Systems, 2020, v. 2020-December-
dc.identifier.issn1049-5258-
dc.identifier.urihttp://hdl.handle.net/10722/327773-
dc.description.abstractIn this paper, we introduce a simplified and unified method for finite-sum convex optimization, named Variance Reduction via Accelerated Dual Averaging (VRADA). In both general convex and strongly convex settings, VRADA can attain an O(n1)-accurate solution in O(n log log n) number of stochastic gradient evaluations which improves the best known result O(n log n), where n is the number of samples. Meanwhile, VRADA matches the lower bound of the general convex setting up to a log log n factor and matches the lower bounds in both regimes n < Θ(K) and n > K of the strongly convex setting, where K denotes the condition number. Besides improving the best known results and matching all the above lower bounds simultaneously, VRADA has more unified and simplified algorithmic implementation and convergence analysis for both the general convex and strongly convex settings. The underlying novel approaches such as the novel initialization strategy in VRADA may be of independent interest. Through experiments on real datasets, we show the good performance of VRADA over existing methods for large-scale machine learning problems.-
dc.languageeng-
dc.relation.ispartofAdvances in Neural Information Processing Systems-
dc.titleVariance reduction via accelerated dual averaging for finite-sum optimization-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.scopuseid_2-s2.0-85108417927-
dc.identifier.volume2020-December-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats