File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Eigen Selection in Spectral Clustering: A Theory-Guided Practice

TitleEigen Selection in Spectral Clustering: A Theory-Guided Practice
Authors
KeywordsAsymptotic expansions
Clustering
Eigen selection
Eigenvalues
Eigenvectors
High dimensionality
Low-rank models
Issue Date2023
Citation
Journal of the American Statistical Association, 2023, v. 118, n. 541, p. 109-121 How to Cite?
AbstractBased on a Gaussian mixture type model of K components, we derive eigen selection procedures that improve the usual spectral clustering algorithms in high-dimensional settings, which typically act on the top few eigenvectors of an affinity matrix (e.g., (Formula presented.)) derived from the data matrix (Formula presented.). Our selection principle formalizes two intuitions: (i) eigenvectors should be dropped when they have no clustering power; (ii) some eigenvectors corresponding to smaller spiked eigenvalues should be dropped due to estimation inaccuracy. Our selection procedures lead to new spectral clustering algorithms: ESSC for K = 2 and GESSC for K > 2. The newly proposed algorithms enjoy better stability and compare favorably against canonical alternatives, as demonstrated in extensive simulation and multiple real data studies. Supplementary materials for this article are available online.
Persistent Identifierhttp://hdl.handle.net/10722/354189
ISSN
2023 Impact Factor: 3.0
2023 SCImago Journal Rankings: 3.922

 

DC FieldValueLanguage
dc.contributor.authorHan, Xiao-
dc.contributor.authorTong, Xin-
dc.contributor.authorFan, Yingying-
dc.date.accessioned2025-02-07T08:47:03Z-
dc.date.available2025-02-07T08:47:03Z-
dc.date.issued2023-
dc.identifier.citationJournal of the American Statistical Association, 2023, v. 118, n. 541, p. 109-121-
dc.identifier.issn0162-1459-
dc.identifier.urihttp://hdl.handle.net/10722/354189-
dc.description.abstractBased on a Gaussian mixture type model of K components, we derive eigen selection procedures that improve the usual spectral clustering algorithms in high-dimensional settings, which typically act on the top few eigenvectors of an affinity matrix (e.g., (Formula presented.)) derived from the data matrix (Formula presented.). Our selection principle formalizes two intuitions: (i) eigenvectors should be dropped when they have no clustering power; (ii) some eigenvectors corresponding to smaller spiked eigenvalues should be dropped due to estimation inaccuracy. Our selection procedures lead to new spectral clustering algorithms: ESSC for K = 2 and GESSC for K > 2. The newly proposed algorithms enjoy better stability and compare favorably against canonical alternatives, as demonstrated in extensive simulation and multiple real data studies. Supplementary materials for this article are available online.-
dc.languageeng-
dc.relation.ispartofJournal of the American Statistical Association-
dc.subjectAsymptotic expansions-
dc.subjectClustering-
dc.subjectEigen selection-
dc.subjectEigenvalues-
dc.subjectEigenvectors-
dc.subjectHigh dimensionality-
dc.subjectLow-rank models-
dc.titleEigen Selection in Spectral Clustering: A Theory-Guided Practice-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1080/01621459.2021.1917418-
dc.identifier.scopuseid_2-s2.0-85107465573-
dc.identifier.volume118-
dc.identifier.issue541-
dc.identifier.spage109-
dc.identifier.epage121-
dc.identifier.eissn1537-274X-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats