File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Calculating optimistic likelihoods using (geodesically) convex optimization

TitleCalculating optimistic likelihoods using (geodesically) convex optimization
Authors
Issue Date2019
Citation
Advances in Neural Information Processing Systems, 2019, v. 32 How to Cite?
AbstractA fundamental problem arising in many areas of machine learning is the evaluation of the likelihood of a given observation under different nominal distributions. Frequently, these nominal distributions are themselves estimated from data, which makes them susceptible to estimation errors. We thus propose to replace each nominal distribution with an ambiguity set containing all distributions in its vicinity and to evaluate an optimistic likelihood, that is, the maximum of the likelihood over all distributions in the ambiguity set. When the proximity of distributions is quantified by the Fisher-Rao distance or the Kullback-Leibler divergence, the emerging optimistic likelihoods can be computed efficiently using either geodesic or standard convex optimization techniques. We showcase the advantages of working with optimistic likelihoods on a classification problem using synthetic as well as empirical data.
Persistent Identifierhttp://hdl.handle.net/10722/313629
ISSN
2020 SCImago Journal Rankings: 1.399
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorNguyen, Viet Anh-
dc.contributor.authorShafieezadeh-Abadeh, Soroosh-
dc.contributor.authorYue, Man Chung-
dc.contributor.authorKuhn, Daniel-
dc.contributor.authorWiesemann, Wolfram-
dc.date.accessioned2022-06-23T01:18:48Z-
dc.date.available2022-06-23T01:18:48Z-
dc.date.issued2019-
dc.identifier.citationAdvances in Neural Information Processing Systems, 2019, v. 32-
dc.identifier.issn1049-5258-
dc.identifier.urihttp://hdl.handle.net/10722/313629-
dc.description.abstractA fundamental problem arising in many areas of machine learning is the evaluation of the likelihood of a given observation under different nominal distributions. Frequently, these nominal distributions are themselves estimated from data, which makes them susceptible to estimation errors. We thus propose to replace each nominal distribution with an ambiguity set containing all distributions in its vicinity and to evaluate an optimistic likelihood, that is, the maximum of the likelihood over all distributions in the ambiguity set. When the proximity of distributions is quantified by the Fisher-Rao distance or the Kullback-Leibler divergence, the emerging optimistic likelihoods can be computed efficiently using either geodesic or standard convex optimization techniques. We showcase the advantages of working with optimistic likelihoods on a classification problem using synthetic as well as empirical data.-
dc.languageeng-
dc.relation.ispartofAdvances in Neural Information Processing Systems-
dc.titleCalculating optimistic likelihoods using (geodesically) convex optimization-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.scopuseid_2-s2.0-85090178488-
dc.identifier.volume32-
dc.identifier.isiWOS:000535866905058-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats