Goto

Collaborating Authors

 stars


Stability Approach to Regularization Selection (StARS) for High Dimensional Graphical Models

Neural Information Processing Systems

A challenging problem in estimating high-dimensional graphical models is to choose the regularization parameter in a data-dependent way. The standard techniques include K -fold cross-validation ( K -CV), Akaike information criterion (AIC), and Bayesian information criterion (BIC). Though these methods work well for low-dimensional problems, they are not suitable in high dimensional settings. The method has a clear interpretation: we use the least amount of regularization that simultaneously makes a graph sparse and replicable under random sampling. This interpretation requires essentially no conditions.


Stability Approach to Regularization Selection (StARS) for High Dimensional Graphical Models

Liu, Han, Roeder, Kathryn, Wasserman, Larry

Neural Information Processing Systems

A challenging problem in estimating high-dimensional graphical models is to choose the regularization parameter in a data-dependent way. The standard techniques include $K$-fold cross-validation ($K$-CV), Akaike information criterion (AIC), and Bayesian information criterion (BIC). Though these methods work well for low-dimensional problems, they are not suitable in high dimensional settings. The method has a clear interpretation: we use the least amount of regularization that simultaneously makes a graph sparse and replicable under random sampling. This interpretation requires essentially no conditions.


Model Selection With Graphical Neighbour Information

O'Shea, Robert

arXiv.org Machine Learning

Accurate m odel selection is a fundamental requirement for statistical analysis (1 - 5) . In many real - world applications of graphical modelling, correct model structure ident ifica tion is the ultimate objective. S tandard model validation procedures such as information theoretic scores and cross validation have demonstr ated poor performance when . Specialised methods such as EBIC, StARS and RIC have been developed for the explicit purpose of high - dimensional Gaussian graphical model selection. We present a novel model score criterion, Graphical Neighbour Information. This method demonstrates oracle performance in high - dimensional model selection, outperforming the current state - of - the - a rt in our simulations. The Graphical Neighbour Information criterion has the additional advantage of efficient, closed - form computability, sparing the costly inference of multiple models on data subsamples. We provide a theoretic analysis of the method and benchmark simulations versus the current state of the art .


Stability Approach to Regularization Selection (StARS) for High Dimensional Graphical Models

Liu, Han, Roeder, Kathryn, Wasserman, Larry

Neural Information Processing Systems

A challenging problem in estimating high-dimensional graphical models is to choose the regularization parameter in a data-dependent way. The standard techniques include $K$-fold cross-validation ($K$-CV), Akaike information criterion (AIC), and Bayesian information criterion (BIC). Though these methods work well for low-dimensional problems, they are not suitable in high dimensional settings. In this paper, we present StARS: a new stability-based method for choosing the regularization parameter in high dimensional inference for undirected graphs. The method has a clear interpretation: we use the least amount of regularization that simultaneously makes a graph sparse and replicable under random sampling. This interpretation requires essentially no conditions. Under mild conditions, we show that StARS is partially sparsistent in terms of graph estimation: i.e. with high probability, all the true edges will be included in the selected model even when the graph size asymptotically increases with the sample size. Empirically, the performance of StARS is compared with the state-of-the-art model selection procedures, including $K$-CV, AIC, and BIC, on both synthetic data and a real microarray dataset. StARS outperforms all competing procedures.


Stability Approach to Regularization Selection (StARS) for High Dimensional Graphical Models

Liu, Han, Roeder, Kathryn, Wasserman, Larry

arXiv.org Machine Learning

A challenging problem in estimating high-dimensional graphical models is to choose the regularization parameter in a data-dependent way. The standard techniques include $K$-fold cross-validation ($K$-CV), Akaike information criterion (AIC), and Bayesian information criterion (BIC). Though these methods work well for low-dimensional problems, they are not suitable in high dimensional settings. In this paper, we present StARS: a new stability-based method for choosing the regularization parameter in high dimensional inference for undirected graphs. The method has a clear interpretation: we use the least amount of regularization that simultaneously makes a graph sparse and replicable under random sampling. This interpretation requires essentially no conditions. Under mild conditions, we show that StARS is partially sparsistent in terms of graph estimation: i.e. with high probability, all the true edges will be included in the selected model even when the graph size diverges with the sample size. Empirically, the performance of StARS is compared with the state-of-the-art model selection procedures, including $K$-CV, AIC, and BIC, on both synthetic data and a real microarray dataset. StARS outperforms all these competing procedures.