File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Neyman-Pearson classification under a strict constraint

TitleNeyman-Pearson classification under a strict constraint
Authors
KeywordsAnomaly detection
Binary classification
Chance constrained optimization
Empirical constraint
Empirical risk minimization
Neyman-Pearson paradigm
Issue Date2011
Citation
Journal of Machine Learning Research, 2011, v. 19, p. 595-613 How to Cite?
AbstractMotivated by problems of anomaly detection, this paper implements the Neyman-Pearson paradigm to deal with asymmetric errors in binary classification with a convex loss. Given a finite collection of classifiers, we combine them and obtain a new classifier that satisfies simultaneously the two following properties with high probability: (i), its probability of type I error is below a pre-specified level and (ii), it has probability of type II error close to the minimum possible. The proposed classifier is obtained by minimizing an empirical objective subject to an empirical constraint. The novelty of the method is that the classifier output by this problem is shown to satisfy the original constraint on type I error. This strict enforcement of the constraint has interesting consequences on the control of the type II error and we develop new techniques to handle this situation. Finally, connections with chance constrained optimization are evident and are investigated. © 2011 P. Rigollet & X. Tong.
Persistent Identifierhttp://hdl.handle.net/10722/354113
ISSN
2023 Impact Factor: 4.3
2023 SCImago Journal Rankings: 2.796

 

DC FieldValueLanguage
dc.contributor.authorRigollet, Philippe-
dc.contributor.authorTong, Xin-
dc.date.accessioned2025-02-07T08:46:33Z-
dc.date.available2025-02-07T08:46:33Z-
dc.date.issued2011-
dc.identifier.citationJournal of Machine Learning Research, 2011, v. 19, p. 595-613-
dc.identifier.issn1532-4435-
dc.identifier.urihttp://hdl.handle.net/10722/354113-
dc.description.abstractMotivated by problems of anomaly detection, this paper implements the Neyman-Pearson paradigm to deal with asymmetric errors in binary classification with a convex loss. Given a finite collection of classifiers, we combine them and obtain a new classifier that satisfies simultaneously the two following properties with high probability: (i), its probability of type I error is below a pre-specified level and (ii), it has probability of type II error close to the minimum possible. The proposed classifier is obtained by minimizing an empirical objective subject to an empirical constraint. The novelty of the method is that the classifier output by this problem is shown to satisfy the original constraint on type I error. This strict enforcement of the constraint has interesting consequences on the control of the type II error and we develop new techniques to handle this situation. Finally, connections with chance constrained optimization are evident and are investigated. © 2011 P. Rigollet & X. Tong.-
dc.languageeng-
dc.relation.ispartofJournal of Machine Learning Research-
dc.subjectAnomaly detection-
dc.subjectBinary classification-
dc.subjectChance constrained optimization-
dc.subjectEmpirical constraint-
dc.subjectEmpirical risk minimization-
dc.subjectNeyman-Pearson paradigm-
dc.titleNeyman-Pearson classification under a strict constraint-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.scopuseid_2-s2.0-84898488452-
dc.identifier.volume19-
dc.identifier.spage595-
dc.identifier.epage613-
dc.identifier.eissn1533-7928-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats