File Download
  Links for fulltext
     (May Require Subscription)
Supplementary

postgraduate thesis: Learning sparse structural associations from data

TitleLearning sparse structural associations from data
Authors
Advisors
Advisor(s):Yiu, SMChin, FYL
Issue Date2017
PublisherThe University of Hong Kong (Pokfulam, Hong Kong)
Citation
Yang, J. [楊杰]. (2017). Learning sparse structural associations from data. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.
AbstractBuilding statistical models to explain the structural association between/among responses (outputs) and predictors (inputs) is important in many real applications, especially for various settings such as bioinformatics, financial prediction and computer vision. A proper choice of the structure which reveals the underline mechanics may be crucial for boosting the performance of prediction models. In this thesis, I propose several methods for automatically learning the latent structural associations for different scenarios. In this thesis I first revisit the problem of multi-response regression and propose an efficient method for feature selection which defined on which we termed the intra- and inter- group sparsities, i.e. for inter-group sparsity I assume only small number of predictors are related to the responses, and for intra-group sparsity each predictor may relate to multiple responses with different sparsity levels. Existing methods fail to model the intra-group sparsity well by either assuming uniform regularization on each group, i.e. each input feature relates to similar number of response features, or requiring prior knowledge of the relationship of predictor and response features. With the new sparsity definition, my method is capable of learning the intra-group sparsity automatically. Second, as responses may not be independent in reality for multi-response prediction, a promising direction is to predict related responses together. However, not all responses have the same degree of relatedness. Sparse Gaussian conditional random field (SGCRF) was developed to learn the degree of relatedness from data without any prior knowledge. However, in real cases, features are not arbitrary, but are dominated by a (smaller) set of related latent factors, e.g. clusters. SGCRF does not capture these latent relations, thus more accurate associations can be explored if we model these latent factors. I propose a novel (mixed membership) hierarchical Bayesian model, namely M2GCRF, to capture this phenomenon in terms of clusters. Third, I explore Bayesian networks (BNs) for modeling conditional distributions of variables and causal relationships. Existing BN structure learning algorithms separately treat those variables with similar tendency. I propose a grouped sparse Gaussian BN (GSGBN) structure learning algorithm which generates the optimal BN with three assumptions: (i) variables follow a multivariate Gaussian distribution, (ii) the network only contains a few edges (sparse), (iii) similar variables have less-divergent sets of parents, while not-so-similar ones should have divergent sets of parents (variable grouping). I make the learned network sparse by L1 regularization, and apply another term to incorporate shared information among variables (Laplacian distributions on the differences of variables’ parents). For similar variables, GSGBN tends to penalize the differences of similar variables’ parent sets more, compared to those not-so-similar variables’ parent sets. With the new definition of the optimal Bayesian network, I obtain the edges of the network and also the similarity of variables from the data, without prior knowledge. Plenty of experiments on both simulated and real data show the above methods have substantially superior prediction performance compared to its competitors in many aspects. (484 words)
DegreeDoctor of Philosophy
SubjectData mining
Sparse matrices
Dept/ProgramComputer Science
Persistent Identifierhttp://hdl.handle.net/10722/261544

 

DC FieldValueLanguage
dc.contributor.advisorYiu, SM-
dc.contributor.advisorChin, FYL-
dc.contributor.authorYang, Jie-
dc.contributor.author楊杰-
dc.date.accessioned2018-09-20T06:44:12Z-
dc.date.available2018-09-20T06:44:12Z-
dc.date.issued2017-
dc.identifier.citationYang, J. [楊杰]. (2017). Learning sparse structural associations from data. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.-
dc.identifier.urihttp://hdl.handle.net/10722/261544-
dc.description.abstractBuilding statistical models to explain the structural association between/among responses (outputs) and predictors (inputs) is important in many real applications, especially for various settings such as bioinformatics, financial prediction and computer vision. A proper choice of the structure which reveals the underline mechanics may be crucial for boosting the performance of prediction models. In this thesis, I propose several methods for automatically learning the latent structural associations for different scenarios. In this thesis I first revisit the problem of multi-response regression and propose an efficient method for feature selection which defined on which we termed the intra- and inter- group sparsities, i.e. for inter-group sparsity I assume only small number of predictors are related to the responses, and for intra-group sparsity each predictor may relate to multiple responses with different sparsity levels. Existing methods fail to model the intra-group sparsity well by either assuming uniform regularization on each group, i.e. each input feature relates to similar number of response features, or requiring prior knowledge of the relationship of predictor and response features. With the new sparsity definition, my method is capable of learning the intra-group sparsity automatically. Second, as responses may not be independent in reality for multi-response prediction, a promising direction is to predict related responses together. However, not all responses have the same degree of relatedness. Sparse Gaussian conditional random field (SGCRF) was developed to learn the degree of relatedness from data without any prior knowledge. However, in real cases, features are not arbitrary, but are dominated by a (smaller) set of related latent factors, e.g. clusters. SGCRF does not capture these latent relations, thus more accurate associations can be explored if we model these latent factors. I propose a novel (mixed membership) hierarchical Bayesian model, namely M2GCRF, to capture this phenomenon in terms of clusters. Third, I explore Bayesian networks (BNs) for modeling conditional distributions of variables and causal relationships. Existing BN structure learning algorithms separately treat those variables with similar tendency. I propose a grouped sparse Gaussian BN (GSGBN) structure learning algorithm which generates the optimal BN with three assumptions: (i) variables follow a multivariate Gaussian distribution, (ii) the network only contains a few edges (sparse), (iii) similar variables have less-divergent sets of parents, while not-so-similar ones should have divergent sets of parents (variable grouping). I make the learned network sparse by L1 regularization, and apply another term to incorporate shared information among variables (Laplacian distributions on the differences of variables’ parents). For similar variables, GSGBN tends to penalize the differences of similar variables’ parent sets more, compared to those not-so-similar variables’ parent sets. With the new definition of the optimal Bayesian network, I obtain the edges of the network and also the similarity of variables from the data, without prior knowledge. Plenty of experiments on both simulated and real data show the above methods have substantially superior prediction performance compared to its competitors in many aspects. (484 words)-
dc.languageeng-
dc.publisherThe University of Hong Kong (Pokfulam, Hong Kong)-
dc.relation.ispartofHKU Theses Online (HKUTO)-
dc.rightsThe author retains all proprietary rights, (such as patent rights) and the right to use in future works.-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subject.lcshData mining-
dc.subject.lcshSparse matrices-
dc.titleLearning sparse structural associations from data-
dc.typePG_Thesis-
dc.description.thesisnameDoctor of Philosophy-
dc.description.thesislevelDoctoral-
dc.description.thesisdisciplineComputer Science-
dc.description.naturepublished_or_final_version-
dc.identifier.doi10.5353/th_991044040578703414-
dc.date.hkucongregation2018-
dc.identifier.mmsid991044040578703414-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats