File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1162/NECO_a_00872
- Scopus: eid_2-s2.0-84988624672
- WOS: WOS:000384452300008
- Find via
Supplementary
- Citations:
- Appears in Collections:
Article: Dimensionality-dependent generalization bounds for k-dimensional coding schemes
Title | Dimensionality-dependent generalization bounds for k-dimensional coding schemes |
---|---|
Authors | |
Issue Date | 2016 |
Citation | Neural Computation, 2016, v. 28, n. 10, p. 2213-2249 How to Cite? |
Abstract | The k-dimensional coding schemes refer to a collection of methods that attempt to represent data using a set of representative k-dimensional vectors and include nonnegative matrix factorization, dictionary learning, sparse coding, k-means clustering, and vector quantization as special cases. Previous generalization bounds for the reconstruction error of the k-dimensional coding schemes are mainly dimensionality-independent. A major advantage of these bounds is that they can be used to analyze the generalization error when data are mapped into an infinite- or high-dimensional feature space. However, many applications use finite-dimensional data features. Can we obtain dimensionality-dependent generalization bounds for k-dimensional coding schemes that are tighter than dimensionality-independent bounds when data are in a finite-dimensional feature space? Yes. In this letter, we address this problem and derive a dimensionality-dependent generalization bound for k-dimensional coding schemes by bounding the covering number of the loss function class induced by the reconstruction error. The bound is of order O((mk ln(mkn)/n)λn, where m is the dimension of features, k is the number of the columns in the linear implementation of coding schemes, and n is the size of sample, λn > 0.5 when n is finite and λn = 0.5 when n is infinite. We show that our bound can be tighter than previous results because it avoids inducing the worst-case upper bound on k of the loss function. The proposed generalization bound is also applied to some specific coding schemes to demonstrate that the dimensionality-dependent bound is an indispensable complement to the dimensionality-independent generalization bounds. |
Persistent Identifier | http://hdl.handle.net/10722/321703 |
ISSN | 2023 Impact Factor: 2.7 2023 SCImago Journal Rankings: 0.948 |
ISI Accession Number ID |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Liu, Tongliang | - |
dc.contributor.author | Tao, Dacheng | - |
dc.contributor.author | Xu, Dong | - |
dc.date.accessioned | 2022-11-03T02:20:53Z | - |
dc.date.available | 2022-11-03T02:20:53Z | - |
dc.date.issued | 2016 | - |
dc.identifier.citation | Neural Computation, 2016, v. 28, n. 10, p. 2213-2249 | - |
dc.identifier.issn | 0899-7667 | - |
dc.identifier.uri | http://hdl.handle.net/10722/321703 | - |
dc.description.abstract | The k-dimensional coding schemes refer to a collection of methods that attempt to represent data using a set of representative k-dimensional vectors and include nonnegative matrix factorization, dictionary learning, sparse coding, k-means clustering, and vector quantization as special cases. Previous generalization bounds for the reconstruction error of the k-dimensional coding schemes are mainly dimensionality-independent. A major advantage of these bounds is that they can be used to analyze the generalization error when data are mapped into an infinite- or high-dimensional feature space. However, many applications use finite-dimensional data features. Can we obtain dimensionality-dependent generalization bounds for k-dimensional coding schemes that are tighter than dimensionality-independent bounds when data are in a finite-dimensional feature space? Yes. In this letter, we address this problem and derive a dimensionality-dependent generalization bound for k-dimensional coding schemes by bounding the covering number of the loss function class induced by the reconstruction error. The bound is of order O((mk ln(mkn)/n)λn, where m is the dimension of features, k is the number of the columns in the linear implementation of coding schemes, and n is the size of sample, λn > 0.5 when n is finite and λn = 0.5 when n is infinite. We show that our bound can be tighter than previous results because it avoids inducing the worst-case upper bound on k of the loss function. The proposed generalization bound is also applied to some specific coding schemes to demonstrate that the dimensionality-dependent bound is an indispensable complement to the dimensionality-independent generalization bounds. | - |
dc.language | eng | - |
dc.relation.ispartof | Neural Computation | - |
dc.title | Dimensionality-dependent generalization bounds for k-dimensional coding schemes | - |
dc.type | Article | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1162/NECO_a_00872 | - |
dc.identifier.scopus | eid_2-s2.0-84988624672 | - |
dc.identifier.volume | 28 | - |
dc.identifier.issue | 10 | - |
dc.identifier.spage | 2213 | - |
dc.identifier.epage | 2249 | - |
dc.identifier.eissn | 1530-888X | - |
dc.identifier.isi | WOS:000384452300008 | - |