File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Dimensionality-dependent generalization bounds for k-dimensional coding schemes

TitleDimensionality-dependent generalization bounds for k-dimensional coding schemes
Authors
Issue Date2016
Citation
Neural Computation, 2016, v. 28, n. 10, p. 2213-2249 How to Cite?
AbstractThe k-dimensional coding schemes refer to a collection of methods that attempt to represent data using a set of representative k-dimensional vectors and include nonnegative matrix factorization, dictionary learning, sparse coding, k-means clustering, and vector quantization as special cases. Previous generalization bounds for the reconstruction error of the k-dimensional coding schemes are mainly dimensionality-independent. A major advantage of these bounds is that they can be used to analyze the generalization error when data are mapped into an infinite- or high-dimensional feature space. However, many applications use finite-dimensional data features. Can we obtain dimensionality-dependent generalization bounds for k-dimensional coding schemes that are tighter than dimensionality-independent bounds when data are in a finite-dimensional feature space? Yes. In this letter, we address this problem and derive a dimensionality-dependent generalization bound for k-dimensional coding schemes by bounding the covering number of the loss function class induced by the reconstruction error. The bound is of order O((mk ln(mkn)/n)λn, where m is the dimension of features, k is the number of the columns in the linear implementation of coding schemes, and n is the size of sample, λn > 0.5 when n is finite and λn = 0.5 when n is infinite. We show that our bound can be tighter than previous results because it avoids inducing the worst-case upper bound on k of the loss function. The proposed generalization bound is also applied to some specific coding schemes to demonstrate that the dimensionality-dependent bound is an indispensable complement to the dimensionality-independent generalization bounds.
Persistent Identifierhttp://hdl.handle.net/10722/321703
ISSN
2023 Impact Factor: 2.7
2023 SCImago Journal Rankings: 0.948
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorLiu, Tongliang-
dc.contributor.authorTao, Dacheng-
dc.contributor.authorXu, Dong-
dc.date.accessioned2022-11-03T02:20:53Z-
dc.date.available2022-11-03T02:20:53Z-
dc.date.issued2016-
dc.identifier.citationNeural Computation, 2016, v. 28, n. 10, p. 2213-2249-
dc.identifier.issn0899-7667-
dc.identifier.urihttp://hdl.handle.net/10722/321703-
dc.description.abstractThe k-dimensional coding schemes refer to a collection of methods that attempt to represent data using a set of representative k-dimensional vectors and include nonnegative matrix factorization, dictionary learning, sparse coding, k-means clustering, and vector quantization as special cases. Previous generalization bounds for the reconstruction error of the k-dimensional coding schemes are mainly dimensionality-independent. A major advantage of these bounds is that they can be used to analyze the generalization error when data are mapped into an infinite- or high-dimensional feature space. However, many applications use finite-dimensional data features. Can we obtain dimensionality-dependent generalization bounds for k-dimensional coding schemes that are tighter than dimensionality-independent bounds when data are in a finite-dimensional feature space? Yes. In this letter, we address this problem and derive a dimensionality-dependent generalization bound for k-dimensional coding schemes by bounding the covering number of the loss function class induced by the reconstruction error. The bound is of order O((mk ln(mkn)/n)λn, where m is the dimension of features, k is the number of the columns in the linear implementation of coding schemes, and n is the size of sample, λn > 0.5 when n is finite and λn = 0.5 when n is infinite. We show that our bound can be tighter than previous results because it avoids inducing the worst-case upper bound on k of the loss function. The proposed generalization bound is also applied to some specific coding schemes to demonstrate that the dimensionality-dependent bound is an indispensable complement to the dimensionality-independent generalization bounds.-
dc.languageeng-
dc.relation.ispartofNeural Computation-
dc.titleDimensionality-dependent generalization bounds for k-dimensional coding schemes-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1162/NECO_a_00872-
dc.identifier.scopuseid_2-s2.0-84988624672-
dc.identifier.volume28-
dc.identifier.issue10-
dc.identifier.spage2213-
dc.identifier.epage2249-
dc.identifier.eissn1530-888X-
dc.identifier.isiWOS:000384452300008-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats