We propose Bayesian methods for Gaussian graphical models that lead to sparse and adaptively shrunk estimators of the precision (inverse covariance) matrix. Our methods are based on lasso‐type regularization priors leading to parsimonious parameterization of the precision matrix, which is essential in several applications involving learning relationships among the variables. In this context, we introduce a novel type of selection prior that develops a sparse structure on the precision matrix by making most of the elements exactly zero, in addition to ensuring positive definiteness—thus conducting model selection and estimation simultaneously. More importantly, we extend these methods to analyse clustered data using finite mixtures of Gaussian graphical model and infinite mixtures of Gaussian graphical models. We discuss appropriate posterior simulation schemes to implement posterior inference in the proposed models, including the evaluation of normalizing constants that are functions of parameters of interest, which result from the restriction of positive definiteness on the correlation matrix. We evaluate the operating characteristics of our method via several simulations and demonstrate the application to real‐data examples in genomics. Copyright © 2014 John Wiley & Sons, Ltd
Bayesian sparse graphical models and their mixtures
R. Talluri,V. Baladandayuthapani,B. Mallick
Published 2013 in Stat
ABSTRACT
PUBLICATION RECORD
- Publication year
2013
- Venue
Stat
- Publication date
2013-10-03
- Fields of study
Mathematics, Computer Science, Medicine
- Identifiers
- External record
- Source metadata
Semantic Scholar, PubMed
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-63 of 63 references · Page 1 of 1
CITED BY
Showing 1-14 of 14 citing papers · Page 1 of 1