File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Classification via minimum incremental coding length (MICL)

TitleClassification via minimum incremental coding length (MICL)
Authors
Issue Date2008
Citation
Advances in Neural Information Processing Systems 20 - Proceedings of the 2007 Conference, 2008 How to Cite?
AbstractWe present a simple new criterion for classification, based on principles from lossy data compression. The criterion assigns a test sample to the class that uses the minimum number of additional bits to code the test sample, subject to an allowable distortion. We prove asymptotic optimality of this criterion for Gaussian data and analyze its relationships to classical classifiers. Theoretical results provide new insights into relationships among popular classifiers such as MAP and RDA, as well as unsupervised clustering methods based on lossy compression [13]. Minimizing the lossy coding length induces a regularization effect which stabilizes the (implicit) density estimate in a small-sample setting. Compression also provides a uniform means of handling classes of varying dimension. This simple classification criterion and its kernel and local versions perform competitively against existing classifiers on both synthetic examples and real imagery data such as handwritten digits and human faces, without requiring domain-specific information.
Persistent Identifierhttp://hdl.handle.net/10722/326622

 

DC FieldValueLanguage
dc.contributor.authorWright, John-
dc.contributor.authorMa, Yi-
dc.contributor.authorTao, Yangyu-
dc.contributor.authorLin, Zhouchen-
dc.contributor.authorShum, Heung Yeung-
dc.date.accessioned2023-03-31T05:25:18Z-
dc.date.available2023-03-31T05:25:18Z-
dc.date.issued2008-
dc.identifier.citationAdvances in Neural Information Processing Systems 20 - Proceedings of the 2007 Conference, 2008-
dc.identifier.urihttp://hdl.handle.net/10722/326622-
dc.description.abstractWe present a simple new criterion for classification, based on principles from lossy data compression. The criterion assigns a test sample to the class that uses the minimum number of additional bits to code the test sample, subject to an allowable distortion. We prove asymptotic optimality of this criterion for Gaussian data and analyze its relationships to classical classifiers. Theoretical results provide new insights into relationships among popular classifiers such as MAP and RDA, as well as unsupervised clustering methods based on lossy compression [13]. Minimizing the lossy coding length induces a regularization effect which stabilizes the (implicit) density estimate in a small-sample setting. Compression also provides a uniform means of handling classes of varying dimension. This simple classification criterion and its kernel and local versions perform competitively against existing classifiers on both synthetic examples and real imagery data such as handwritten digits and human faces, without requiring domain-specific information.-
dc.languageeng-
dc.relation.ispartofAdvances in Neural Information Processing Systems 20 - Proceedings of the 2007 Conference-
dc.titleClassification via minimum incremental coding length (MICL)-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.scopuseid_2-s2.0-85138462036-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats