Timezone: »
We analyze the complexity of learning directed acyclic graphical models from observational data in general settings without specific distributional assumptions. Our approach is information-theoretic and uses a local Markov boundary search procedure in order to recursively construct ancestral sets in the underlying graphical model. Perhaps surprisingly, we show that for certain graph ensembles, a simple forward greedy search algorithm (i.e. without a backward pruning phase) suffices to learn the Markov boundary of each node. This substantially improves the sample complexity, which we show is at most polynomial in the number of nodes. This is then applied to learn the entire graph under a novel identifiability condition that generalizes existing conditions from the literature. As a matter of independent interest, we establish finite-sample guarantees for the problem of recovering Markov boundaries from data. Moreover, we apply our results to the special case of polytrees, for which the assumptions simplify, and provide explicit conditions under which polytrees are identifiable and learnable in polynomial time. We further illustrate the performance of the algorithm, which is easy to implement, in a simulation study. Our approach is general, works for discrete or continuous distributions without distributional assumptions, and as such sheds light on the minimal assumptions required to efficiently learn the structure of directed graphical models from data.
Author Information
Ming Gao (the University of Chicago)
Bryon Aragam (University of Chicago)
More from the Same Authors
-
2022 Spotlight: Identifiability of deep generative models without auxiliary information »
Bohdan Kivva · Goutham Rajendran · Pradeep Ravikumar · Bryon Aragam -
2022 Poster: DAGMA: Learning DAGs via M-matrices and a Log-Determinant Acyclicity Characterization »
Kevin Bello · Bryon Aragam · Pradeep Ravikumar -
2022 Poster: Identifiability of deep generative models without auxiliary information »
Bohdan Kivva · Goutham Rajendran · Pradeep Ravikumar · Bryon Aragam -
2021 Poster: Learning latent causal graphs via mixture oracles »
Bohdan Kivva · Goutham Rajendran · Pradeep Ravikumar · Bryon Aragam -
2021 Poster: Structure learning in polynomial time: Greedy algorithms, Bregman information, and exponential families »
Goutham Rajendran · Bohdan Kivva · Ming Gao · Bryon Aragam -
2020 Poster: A polynomial-time algorithm for learning nonparametric causal graphs »
Ming Gao · Yi Ding · Bryon Aragam -
2019 Poster: Learning Sample-Specific Models with Low-Rank Personalized Regression »
Ben Lengerich · Bryon Aragam · Eric Xing -
2019 Poster: Globally optimal score-based learning of directed acyclic graphs in high-dimensions »
Bryon Aragam · Arash Amini · Qing Zhou -
2018 Poster: The Sample Complexity of Semi-Supervised Learning with Nonparametric Mixture Models »
Chen Dan · Liu Leqi · Bryon Aragam · Pradeep Ravikumar · Eric Xing -
2018 Poster: DAGs with NO TEARS: Continuous Optimization for Structure Learning »
Xun Zheng · Bryon Aragam · Pradeep Ravikumar · Eric Xing -
2018 Spotlight: DAGs with NO TEARS: Continuous Optimization for Structure Learning »
Xun Zheng · Bryon Aragam · Pradeep Ravikumar · Eric Xing