// about

**statistical machine learning** | nonparametric statistics, minimax rates, unsupervised learning, deep learning, high-dimensional statistics

**latent variable models** | deep generative models, probabilistic models, representation learning, mixture models, variational inference

**graphical models and causality** | structural equation models, model selection, structure learning, causal models

// selected papers

Uniform consistency in nonparametric mixture models, B Aragam and R Yang (Annals of Statistics)

Tight bounds on the hardness of learning simple nonparametric mixtures, WM Tai and B Aragam (COLT)

Identifiability of deep generative models without auxiliary information, B Kivva, G Rajendran, P Ravikumar and B Aragam (NeurIPS)

Identifiability of nonparametric mixture models and Bayes optimal clustering, B Aragam, C Dan, EP Xing, and P Ravikumar (Annals of Statistics)

Optimal estimation of Gaussian DAG models, M Gao, WM Tai, and B Aragam (AISTATS)

Learning latent causal graphs via mixture oracles, B Kivva, G Rajendran, P Ravikumar and B Aragam (NeurIPS)

A polynomial-time algorithm for learning nonparametric causal graphs, M Gao, Y Ding, and B Aragam (NeurIPS)

DAGs with NO TEARS: Continuous optimization for structure learning, X Zheng, B Aragam, P Ravikumar, and EP Xing (NeurIPS, spotlight)

// news

- Several papers accepted to NeurIPS, including an oral presentation.
- New preprint: How difficult is neighbourhood selection? New results on minimax optimality and the effect of path cancellation (unfaithfulness).
- Several new preprints on causal representation learning and nonparametric latent variable models: Learning latent causal graphs, learning nonlinear causal representations, and neuro-causal factor analysis.
- New preprint: Optimization theory for NOTEARS objectives and gradient-based DAG learning (Update: Accepted to ICML 2023).
- Our paper on the hardness of learning nonparametric mixture models has been accepted to COLT 2023.
- New preprint: Sample complexity bounds for learning mixtures with censored data (Update: Accepted to ICML 2023).
- Two papers accepted to ICML.
- Our paper on uniform consistency in estimating nonparametric mixture models has been accepted to the Annals of Statistics.
- Code for the DAGMA algorithm is now available on GitHub: Continuous optimization for structure learning with a faster and more accurate log-det constraint.
- Two papers accepted to NeurIPS.
- New preprint: Faster and more accurate DAG learning through a new acyclicity characterization based on log-det constraints and M-matrices (Update: Accepted to NeurIPS 2022).
- Our paper on invariant representation learning and its fundamental limits has been accepted to JMLR.
- New preprint: Learning nonparametric latent variable models is provably hard: Tight, super-polynomial lower bounds on the sample complexity of this problem (Update: Accepted to COLT 2023).
- New preprint: New nonparametric identifiability results for deep generative models such as VAEs that are applicable to commonly used architectures (Update: Accepted to NeurIPS 2022).
- New preprint: A new way to encode conditional independence without faithfulness using a lattice-theoretic approach.
- Two papers accepted to AISTATS.
- New preprint: The first optimal sample complexity bounds for learning directed acyclic graphical models (Update: Accepted to AISTATS 2022).
- Three papers accepted to NeurIPS.
- New preprint: A general, distribution-free approach to identifying and learning directed acyclic graphical models (Update: Accepted to NeurIPS 2021).
- New preprint: Connections between score-based learning, greedy forward-backward search, and existing polynomial-time algorithms for structure learning (Update: Accepted to NeurIPS 2021).
- New preprint: New identifiability results and algorithms for learning causal models with latent variables (Update: Accepted to NeurIPS 2021).
- New preprint: A detailed study of uniform consistency in estimating nonparametric mixture models.
- NSF grant on learning nonparametric causal models has been funded.
- NIH grant on learning personalized models has been funded.
- New preprint: A general approach to analyzing upper and lower bounds in representation learning for fairness, privacy, and related problems.
- New preprint: A polynomial-time algorithm for learning nonparametric causal graphs with finite-sample guarantees.
- Our paper on automating dependence plots for explainable machine learning was accepted to UAI 2020.
- The code for extending NOTEARS to nonparametric models is now available on GitHub. See this paper and this blog post for more details.
- Two papers accepted to AISTATS 2020.
- New preprint: Learning structural VAR models from time series data via black-box optimization.
- New preprint: A new approach to personalization via interpretable, sample-specific models.
- New preprint: A general framework for score-based learning of nonparametric DAG models.
- Two papers accepted to NeurIPS 2019.
- New preprint: A new proof that almost all Gaussian graphical models are perfect.
- Our paper on nonparametric mixture models and clustering has been accepted to the Annals of Statistics.
- Our paper on fault tolerance in machine learning has been accepted to ICML 2019.
- Code for the NO TEARS algorithm is now available on GitHub. Structure learning in less than 50 lines of code!
- Preprint posted: A general framework for quantifying the cost of system faults during ML training.
- Two papers accepted to NeurIPS 2018.
- Preprint posted: New lower bounds on the labeled sample complexity of semi-supervised learning without parametric assumptions.
- Code for (a) the Precision Lasso and (b) personalized regression is now available on GitHub.
- New paper accepted to Bioinformatics.
- New paper accepted to ISMB 2018, to appear in Bioinformatics.
- New release of sparsebn package for R: Support for white and blacklists, Cytoscape, and more!
- Preprint posted: A smooth relaxation of DAG learning, solved with black-box optimizers.
- Preprint posted: New theory on the identifiability of nonparametric mixture models with applications to clustering.
- Our paper on the sparsebn package for R is accepted to JSS.

// publications

// papers

*alphabetical order

click to:
filter by topic
/
filter by venue

Uncovering meanings of embeddings via Markov boundaries

Y Jiang, B Aragam, V Veitch

NeurIPS 2023, to appear

preprint

Embeddings like CLIP, GloVe, and word2vec are an integral part of large-scale machine-learning models. There is evidence that embedding methods encode high-level semantic information into the vector space structure of the embedding space. In this paper, we study the role of partial orthogonality in encoding meanings by searching for ``meaningful'' subspaces of an embedding spanned by other embeddings, which generalizes the notion of Markov boundaries in Euclidean space. Using this tool, we empirically study the semantic meaning of partial orthogonality in CLIP embeddings and find a good match to conceptual semantic meaning. Complementary to this, we also introduce the concept of independence preserving embeddings where embeddings preserve the conditional independence structures of a distribution, and we prove the existence of such embeddings and approximations to them.

Keywords: machine learning, embeddings, semantic meaning, Markov boundary, partial orthogonality

Global optimality in bivariate gradient-based DAG learning

C Deng, K Bello, B Aragam, and P Ravikumar

NeurIPS 2023, to appear

preprint

Recently, a new class of non-convex optimization problems motivated by the statistical problem of learning an acyclic directed graphical model from data has attracted significant interest. While existing work uses standard first-order optimization schemes to solve this problem, proving the global optimality of such approaches has proven elusive. The difficulty lies in the fact that unlike other non-convex problems in the literature, this problem is not "benign", and possesses multiple spurious solutions that standard approaches can easily get trapped in. In this paper, we prove that a simple path-following optimization scheme globally converges to the global minimum of the population loss in the bivariate setting.

Keywords: nonconvex optimization, global optimization, homotopy, graphical models, directed acyclic graphs

iSCAN: Identifying causal mechanism shifts among nonlinear additive noise models

T Chen, K Bello, B Aragam, and P Ravikumar

NeurIPS 2023, to appear

preprint

Structural causal models (SCMs) are widely used in various disciplines to represent causal relationships among variables in complex systems. Unfortunately, the true underlying directed acyclic graph (DAG) structure is often unknown, and determining it from observational or interventional data remains a challenging task. However, in many situations, the end goal is to identify changes (shifts) in causal mechanisms between related SCMs rather than recovering the entire underlying DAG structure. Examples include analyzing gene regulatory network structure changes between healthy and cancerous individuals or understanding variations in biological pathways under different cellular contexts. This paper focuses on identifying functional mechanism shifts in two or more related SCMs over the same set of variables -- without estimating the entire DAG structure of each SCM. Prior work under this setting assumed linear models with Gaussian noises; instead, in this work we assume that each SCM belongs to the more general class of nonlinear additive noise models (ANMs). A key contribution of this work is to show that the Jacobian of the score function for the mixture distribution allows for identification of shifts in general non-parametric functional mechanisms. Once the shifted variables are identified, we leverage recent work to estimate the structural differences, if any, for the shifted variables. Experiments on synthetic and real-world data are provided to showcase the applicability of this approach.

Keywords: graphical models, directed acyclic graphs, causality, causal representation learning, mechanism shift

Optimal neighbourhood selection in structural equation models

M Gai, WM Tai, and B Aragam

preprint

We study the optimal sample complexity of neighbourhood selection in linear structural equation models, and compare this to best subset selection (BSS) for linear models under general design. We show by example that -- even when the structure is \emph{unknown} -- the existence of underlying structure can reduce the sample complexity of neighbourhood selection. This result is complicated by the possibility of path cancellation, which we study in detail, and show that improvements are still possible in the presence of path cancellation. Finally, we support these theoretical observations with experiments. The proof introduces a modified BSS estimator, called klBSS, and compares its performance to BSS. The analysis of klBSS may also be of independent interest since it applies to arbitrary structured models, not necessarily those induced by a structural equation model. Our results have implications for structure learning in graphical models, which often relies on neighbourhood selection as a subroutine.

Keywords: neighbourhood selection, support recovery, structural equations, graphical models

Learning nonparametric latent causal graphs with unknown interventions

Y Jiang and B Aragam

NeurIPS 2023, to appear

preprint

We establish conditions under which latent causal graphs are nonparametrically identifiable and can be reconstructed from unknown interventions in the latent space. Our primary focus is the identification of the latent structure in a measurement model, i.e. causal graphical models where dependence between observed variables is insignificant compared to dependence between latent representations, without making parametric assumptions such as linearity or Gaussianity. Moreover, we do not assume the number of hidden variables is known, and we show that at most one unknown intervention per hidden variable is needed. This extends a recent line of work on learning causal representations from observations and interventions. The proofs are constructive and introduce two new graphical concepts -- imaginary subsets and isolated edges -- that may be useful in their own right. As a matter of independent interest, the proofs also involve a novel characterization of the limits of edge orientations within the equivalence class of DAGs induced by unknown interventions. Experiments confirm that the latent graph can be recovered from data using our theoretical results. These are the first results to characterize the conditions under which causal representations are identifiable without making any parametric assumptions in a general setting with unknown interventions and without faithfulness.

Keywords: graphical models, directed acyclic graphs, causality, identifiability, causal representation learning, unknown interventions

Learning linear causal representations from interventions under general nonlinear mixing

S Buchholz, G Rajendran, E Rosenfeld, B Aragam, B Schölkopf, P Ravikumar

NeurIPS 2023, to appear (oral)

preprint

We study the problem of learning causal representations from unknown, latent interventions in a general setting, where the latent distribution is Gaussian but the mixing function is completely general. We prove strong identifiability results given unknown single-node interventions, i.e., without having access to the intervention targets. This generalizes prior works which have focused on weaker classes, such as linear maps or paired counterfactual data. This is also the first instance of causal identifiability from non-paired interventions for deep neural network embeddings. Our proof relies on carefully uncovering the high-dimensional geometric structure present in the data distribution after a non-linear density transformation, which we capture by analyzing quadratic forms of precision matrices of the latent distributions. Finally, we propose a contrastive algorithm to identify the latent variables in practice and evaluate its performance on various tasks.

Keywords: causal representation learning, interventions, experimental data, structural causal models

Neuro-causal factor analysis

A Markham, M Liu, B Aragam, and L Solus

preprint

Factor analysis (FA) is a statistical tool for studying how observed variables with some mutual dependences can be expressed as functions of mutually independent unobserved factors, and it is widely applied throughout the psychological, biological, and physical sciences. We revisit this classic method from the comparatively new perspective given by advancements in causal discovery and deep learning, introducing a framework for Neuro-Causal Factor Analysis (NCFA). Our approach is fully nonparametric: it identifies factors via latent causal discovery methods and then uses a variational autoencoder (VAE) that is constrained to abide by the Markov factorization of the distribution with respect to the learned graph. We evaluate NCFA on real and synthetic data sets, finding that it performs comparably to standard VAEs on data reconstruction tasks but with the advantages of sparser architecture, lower model complexity, and causal interpretability. Unlike traditional FA methods, our proposed NCFA method allows learning and reasoning about the latent factors underlying observed data from a justifiably causal perspective, even when the relations between factors and measurements are highly nonlinear.

Keywords: factor analysis, causal discovery, variational autoencoder, latent variable models, generative models, causality

Optimizing NOTEARS objectives via topological swaps

C Deng, K Bello, B Aragam and P Ravikumar

ICML 2023

preprint

Recently, an intriguing class of non-convex optimization problems has emerged in the context of learning directed acyclic graphs (DAGs). These problems involve minimizing a given loss or score function, subject to a non-convex continuous constraint that penalizes the presence of cycles in a graph. In this work, we delve into the optimality challenges associated with this class of non-convex programs. To address these challenges, we propose a bi-level algorithm that leverages the non-convex constraint in a novel way. The outer level of the algorithm optimizes over topological orders by iteratively swapping pairs of nodes within the topological order of a DAG. A key innovation of our approach is the development of an effective method for generating a set of candidate swapping pairs for each iteration. At the inner level, given a topological order, we utilize off-the-shelf solvers that can handle linear constraints. The key advantage of our proposed algorithm is that it is guaranteed to find a local minimum or a KKT point under weaker conditions compared to previous work and finds solutions with lower scores. Extensive experiments demonstrate that our method outperforms state-of-the-art approaches in terms of achieving a better score. Additionally, our method can also be used as a post-processing algorithm to significantly improve the score of other algorithms.

Keywords: nonconvex optimization, nonlinear constraints, KKT conditions, directed acyclic graphs, Bayesian networks

Learning mixtures of gaussians with censored data

*B Aragam and WM Tai

ICML 2023

preprint

We study the problem of learning mixtures of Gaussians with censored data. Statistical learning with censored data is a classical problem, with numerous practical applications, however, finite-sample guarantees for even simple latent variable models such as Gaussian mixtures are missing. Formally, we are given censored data from a mixture of univariate Gaussians $$\sum_{i=1}^k w_i \mathcal{N}(\mu_i,\sigma^2),$$ i.e. the sample is observed only if it lies inside a set $S$. The goal is to learn the weights $w_i$ and the means $\mu_i$. We propose an algorithm that takes only $\frac{1}{\varepsilon^{O(k)}}$ samples to estimate the weights $w_i$ and the means $\mu_i$ within $\varepsilon$ error.

Keywords: learning theory, mixture models, sample complexity, censored data

Tight bounds on the hardness of learning simple nonparametric mixtures

*B Aragam and WM Tai

COLT 2023

preprint

We study the problem of learning nonparametric distributions in a finite mixture, and establish tight bounds on the sample complexity for learning the component distributions in such models. Namely, we are given i.i.d. samples from a pdf $f$ where $$f=\sum_{i=1}^k w_i f_i, \quad\sum_{i=1}^k w_i=1, \quad w_i>0$$ and we are interested in learning each component $f_i$. Without any assumptions on $f_i$, this problem is ill-posed. In order to identify the components $f_i$, we assume that each $f_i$ can be written as a convolution of a Gaussian and a compactly supported density $\nu_i$ with $\text{supp}(\nu_i)\cap \text{supp}(\nu_j)=\emptyset$.

Our main result shows that $(\frac{1}{\varepsilon})^{\Omega(\log\log \frac{1}{\varepsilon})}$ samples are required for estimating each $f_i$. Unlike parametric mixtures, the difficulty does not arise from the order $k$ or small weights $w_i$, and unlike nonparametric density estimation it does not arise from the curse of dimensionality, irregularity, or inhomogeneity. The proof relies on a fast rate for approximation with Gaussians, which may be of independent interest. To show this is tight, we also propose an algorithm that uses $(\frac{1}{\varepsilon})^{O(\log\log \frac{1}{\varepsilon})}$ samples to estimate each $f_i$. Unlike existing approaches to learning latent variable models based on moment-matching and tensor methods, our proof instead involves a delicate analysis of an ill-conditioned linear system via orthogonal functions. Combining these bounds, we conclude that the optimal sample complexity of this problem properly lies in between polynomial and exponential.

Keywords: learning theory, nonparametric statistics, sample complexity, lower bounds, deconvolution

DAGMA: Learning DAGs via M-matrices and a Log-Determinant Acyclicity Characterization

K Bello, B Aragam and P Ravikumar

NeurIPS 2022

proceedings /
preprint

The combinatorial problem of learning directed acyclic graphs (DAGs) from data was recently framed as a purely continuous optimization problem by leveraging a differentiable acyclicity characterization of DAGs based on the trace of a matrix exponential function. Existing acyclicity characterizations are based on the idea that powers of an adjacency matrix contain information about walks and cycles. In this work, we propose a new acyclicity characterization based on the log-determinant (log-det) function, which leverages the nilpotency property of DAGs. To deal with the inherent asymmetries of a DAG, we relate the domain of our log-det characterization to the set of $\textit{M-matrices}$, which is a key difference to the classical log-det function defined over the cone of positive definite matrices. Similar to acyclicity functions previously proposed, our characterization is also exact and differentiable. However, when compared to existing characterizations, our log-det function: (1) Is better at detecting large cycles; (2) Has better-behaved gradients; and (3) Its runtime is in practice about an order of magnitude faster. From the optimization side, we drop the typically used augmented Lagrangian scheme and propose DAGMA ($\textit{Directed Acyclic Graphs via M-matrices for Acyclicity}$), a method that resembles the central path for barrier methods. Each point in the central path of DAGMA is a solution to an unconstrained problem regularized by our log-det function, then we show that at the limit of the central path the solution is guaranteed to be a DAG. Finally, we provide extensive experiments for $\textit{linear}$ and $\textit{nonlinear}$ SEMs and show that our approach can reach large speed-ups and smaller structural Hamming distances against state-of-the-art methods.

Keywords:structure learning, causal discovery, graphical models, continuous optimization, directed acyclic graphs

Identifiability of deep generative models without auxiliary information

B Kivva, G Rajendran, P Ravikumar and B Aragam

NeurIPS 2022

proceedings /
preprint

We prove identifiability of a broad class of deep latent variable models that (a) have universal approximation capabilities and (b) are the decoders of variational autoencoders that are commonly used in practice. Unlike existing work, our analysis does not require weak supervision, auxiliary information, or conditioning in the latent space. Recently, there has been a surge of works studying identifiability of such models. In these works, the main assumption is that along with the data, an auxiliary variable u (also known as side information) is observed as well. At the same time, several works have empirically observed that this doesn't seem to be necessary in practice. In this work, we explain this behavior by showing that for a broad class of generative (i.e. unsupervised) models with universal approximation capabilities, the side information u is not necessary: We prove identifiability of the entire generative model where we do not observe u and only observe the data x. The models we consider are tightly connected with autoencoder architectures used in practice that leverage mixture priors in the latent space and ReLU/leaky-ReLU activations in the encoder. Our main result is an identifiability hierarchy that significantly generalizes previous work and exposes how different assumptions lead to different "strengths" of identifiability. For example, our weakest result establishes (unsupervised) identifiability up to an affine transformation, which already improves existing work. It's well known that these models have universal approximation capabilities and moreover, they have been extensively used in practice to learn representations of data.

Keywords: deep generative models, identifiability, nonlinear ICA, variational autoencoder, ReLU activations

A non-graphical representation of conditional independence via the neighbourhood lattice

*AA Amini, B Aragam and Q Zhou

preprint

We introduce and study the neighbourhood lattice decomposition of a distribution, which is a compact, non-graphical representation of conditional independence that is valid in the absence of a faithful graphical representation. The idea is to view the set of neighbourhoods of a variable as a subset lattice, and partition this lattice into convex sublattices, each of which directly encodes a collection of conditional independence relations. We show that this decomposition exists in any compositional graphoid and can be computed efficiently and consistently in high-dimensions. In particular, this gives a way to encode all of independence relations implied by a distribution that satisfies the composition axiom, which is strictly weaker than the faithfulness assumption that is typically assumed by graphical approaches. We also discuss various special cases such as graphical models and projection lattices, each of which has intuitive interpretations. Along the way, we see how this problem is closely related to neighbourhood regression, which has been extensively studied in the context of graphical models and structural equations.

Keywords: conditional independence, graphical models, neighbourhood lattice, computation, compositional graphoids

Optimal estimation of Gaussian DAG models

M Gao, WM Tai, and B Aragam

AISTATS 2022

proceedings /
preprint

We study the optimal sample complexity of learning a Gaussian directed acyclic graph (DAG) from observational data. Our main results establish the minimax optimal sample complexity for learning the structure of a linear Gaussian DAG model in two settings of interest: 1) Under equal variances without knowledge of the true ordering, and 2) For general linear models given knowledge of the ordering. In both cases the sample complexity is $n\asymp q\log(d/q)$, where $q$ is the maximum number of parents and $d$ is the number of nodes. We further make comparisons with the classical problem of learning (undirected) Gaussian graphical models, showing that under the equal variance assumption, these two problems share the same optimal sample complexity. In other words, at least for Gaussian models with equal error variances, learning a directed graphical model is statistically no more difficult than learning an undirected graphical model. Our results also extend to more general identification assumptions as well as subgaussian errors.

Keywords: graphical models, sample complexity, lower bounds, equal variances

On perfectness in Gaussian graphical models

AA Amini, B Aragam, and Q Zhou

AISTATS 2022

proceedings /
preprint /
code

Knowing when a graphical model is perfect to a distribution is essential in order to relate separation in the graph to conditional independence in the distribution, and this is particularly important when performing inference from data. When the model is perfect, there is a one-to-one correspondence between conditional independence statements in the distribution and separation statements in the graph. Previous work has shown that almost all models based on linear directed acyclic graphs as well as Gaussian chain graphs are perfect, the latter of which subsumes Gaussian graphical models (i.e., the undirected Gaussian models) as a special case. However, the complexity of chain graph models leads to a proof of this result which is indirect and mired by the complications of parameterizing this general class. In this paper, we directly approach the problem of perfectness for the Gaussian graphical models, and provide a new proof, via a more transparent parametrization, that almost all such models are perfect. Our approach is based on, and substantially extends, a construction of Lněnička and Matúš showing the existence of a perfect Gaussian distribution for any graph.

Keywords: graphical models, perfectness, conditional independence graphs

Efficient Bayesian network structure learning via local Markov boundary search

M Gao and B Aragam

NeurIPS 2021

proceedings /
preprint /
code

We analyze the complexity of learning directed acyclic graphical models from observational data in general settings without specific distributional assumptions. Our approach is information-theoretic and uses a local Markov boundary search procedure in order to recursively construct ancestral sets in the underlying graphical model. Perhaps surprisingly, we show that for certain graph ensembles, a simple forward greedy search algorithm (i.e. without a backward pruning phase) suffices to learn the Markov boundary of each node. This substantially improves the sample complexity, which we show is at most polynomial in the number of nodes. This is then applied to learn the entire graph under a novel identifiability condition that generalizes existing conditions from the literature. As a matter of independent interest, we establish finite-sample guarantees for the problem of recovering Markov boundaries from data. Moreover, we apply our results to the special case of polytrees, for which the assumptions simplify, and provide explicit conditions under which polytrees are identifiable and learnable in polynomial time. We further illustrate the performance of the algorithm, which is easy to implement, in a simulation study. Our approach is general, works for discrete or continuous distributions without distributional assumptions, and as such sheds light on the minimal assumptions required to efficiently learn the structure of directed graphical models from data.

Keywords: Bayesian networks, structure learning, Markov boundary, sample complexity

Structure learning in polynomial time: Greedy algorithms, Bregman information, and exponential families

G Rajendran, B Kivva, M Gao and B Aragam

NeurIPS 2021

proceedings /
preprint

Greedy algorithms have long been a workhorse for learning graphical models, and more broadly for learning statistical models with sparse structure. In the context of learning directed acyclic graphs, greedy algorithms are popular despite their worst-case exponential runtime. In practice, however, they are very efficient. We provide new insight into this phenomenon by studying a general greedy score-based algorithm for learning DAGs. Unlike edge-greedy algorithms such as the popular GES and hill-climbing algorithms, our approach is vertex-greedy and requires at most a polynomial number of score evaluations. We then show how recent polynomial-time algorithms for learning DAG models are a special case of this algorithm, thereby illustrating how these order-based algorithms can be rigourously interpreted as score-based algorithms. This observation suggests new score functions and optimality conditions based on the duality between Bregman divergences and exponential families, which we explore in detail. Explicit sample and computational complexity bounds are derived. Finally, we provide extensive experiments suggesting that this algorithm indeed optimizes the score in a variety of settings.

Keywords: directed acyclic graphs, structure learning, greedy algorithms, Bregman divergence, Bregman information, exponential families

Learning latent causal graphs via mixture oracles

B Kivva, G Rajendran, P Ravikumar and B Aragam

NeurIPS 2021

proceedings /
preprint

We study the problem of reconstructing a causal graphical model from data in the presence of latent variables. The main problem of interest is recovering the causal structure over the latent variables while allowing for general, potentially nonlinear dependence between the variables. In many practical problems, the dependence between raw observations (e.g. pixels in an image) is much less relevant than the dependence between certain high-level, latent features (e.g. concepts or objects), and this is the setting of interest. We provide conditions under which both the latent representations and the underlying latent causal model are identifiable by a reduction to a mixture oracle. The proof is constructive, and leads to several algorithms for explicitly reconstructing the full graphical model. We discuss efficient algorithms and provide experiments illustrating the algorithms in practice.

Keywords: directed acyclic graphs, latent variable models, algorithms, identifiability, causal inference

Uniform consistency in nonparametric mixture models

*B Aragam and R Yang

Annals of Statistics

journal /
preprint

We study uniform consistency in nonparametric mixture models as well as closely related mixture of regression (also known as mixed regression) models, where the regression functions are allowed to be nonparametric and the error distributions are assumed to be convolutions of a Gaussian density. We construct uniformly consistent estimators under general conditions while simultaneously highlighting several pain points in extending existing pointwise consistency results to uniform results. The resulting analysis turns out to be nontrivial, and several novel technical tools are developed along the way. In the case of mixed regression, we prove $L^1$ convergence of the regression functions while allowing for the component regression functions to intersect arbitrarily often, which presents additional technical challenges. We also consider generalizations to general (i.e. non-convolutional) nonparametric mixtures.

Keywords: mixture models, mixed regression, nonparametric estimation, uniform consistency

Fundamental limits and tradeoffs in invariant representation learning

H Zhao, C Dan, B Aragam, T Jaakkola, G Gordon, and P Ravikumar

Journal of Machine Learning Research

journal /
preprint

Many machine learning applications involve learning representations that achieve two competing goals: To maximize information or accuracy with respect to a subset of features (e.g. for prediction) while simultaneously maximizing invariance or independence with respect to another, potentially overlapping, subset of features (e.g. for fairness). Typical examples include privacy-preserving learning, domain adaptation, and algorithmic fairness, just to name a few. In fact, all of the above problems admit a common minimax game-theoretic formulation, whose equilibrium represents a fundamental tradeoff between accuracy and invariance.

In this paper, we provide an information theoretic analysis of this general and important problem under both classification and regression settings. In both cases, we analyze the inherent tradeoffs between accuracy and invariance by providing a geometric characterization of the feasible region in the information plane, where we connect the geometric properties of this feasible region to the fundamental limitations of the tradeoff problem. In the regression setting, we also derive a tight lower bound on the Lagrangian objective that quantifies the tradeoff between accuracy and invariance. This lower bound leads to a better understanding of the tradeoff via the spectral properties of the joint distribution. In both cases, our results shed new light on this fundamental problem by providing insights on the interplay between accuracy and invariance. These results deepen our understanding of this fundamental problem and may be useful in guiding the design of adversarial representation learning algorithms.

Keywords: representation learning, information theory, invariance, fairness, lower bounds

A polynomial-time algorithm for learning nonparametric causal graphs

M Gao, Y Ding, and B Aragam

NeurIPS 2020

proceedings /
preprint /
code

We establish finite-sample guarantees for a polynomial-time algorithm for learning a nonlinear, nonparametric directed acyclic graphical (DAG) model from data. The analysis is model-free and does not assume linearity, additivity, independent noise, or faithfulness. Instead, we impose a condition on the residual variances that is closely related to previous work on linear models with equal variances. Compared to an optimal algorithm with oracle knowledge of the variable ordering, the additional cost of the algorithm is linear in the dimension $d$ and the number of samples $n$. Finally, we compare the proposed algorithm to existing approaches in a simulation study.

Keywords: directed acyclic graphs, nonparametric statistics, algorithms, causality

Automated dependence plots

D Inouye, L Leqi, JS Kim, B Aragam, and P Ravikumar

UAI 2020

proceedings /
preprint /
code

In practical applications of machine learning, it is necessary to look beyond standard metrics such as test accuracy in order to validate various qualitative properties of a model. Partial dependence plots (PDP), including instance-specific PDPs (i.e., ICE plots), have been widely used as a visual tool to understand or validate a model. Yet, current PDPs suffer from two main drawbacks: (1) a user must manually sort or select interesting plots, and (2) PDPs are usually limited to plots along a single feature. To address these drawbacks, we formalize a method for automating the selection of interesting PDPs and extend PDPs beyond showing single features to show the model response along arbitrary directions, for example in raw feature space or a latent space arising from some generative model. We demonstrate the usefulness of our automated dependence plots (ADP) across multiple use-cases and datasets including model selection, bias detection, understanding out-of-sample behavior, and exploring the latent space of a generative model.

Keywords: model explanations, dependency plots, fairness, diagnostics, validation

Learning sparse nonparametric DAGs

X Zheng, C Dan, B Aragam, P Ravikumar, and EP Xing

AISTATS 2020

proceedings /
preprint /
code /
blog

We develop a framework for learning sparse nonparametric directed acyclic graphs (DAGs) from data. Our approach is based on a recent algebraic characterization of DAGs that led to the first fully continuous optimization for score-based learning of DAG models parametrized by a linear structural equation model (SEM). We extend this algebraic characterization to nonparametric SEM by leveraging nonparametric sparsity based on partial derivatives, resulting in a continuous optimization problem that can be applied to a variety of nonparametric and semiparametric models including GLMs, additive noise models, and index models as special cases. We also explore the use of neural networks and orthogonal basis expansions to model nonlinearities for general nonparametric models. Extensive empirical study confirms the necessity of nonlinear dependency and the advantage of continuous optimization for score-based learning.

Keywords: directed acyclic graphs, Bayesian networks, nonparametric statistics, multilayer perceptron, basis expansions

DYNOTEARS: Structure learning from time-series data

R Pamfil,
N Sriwattanaworachai,
P Pilgerstorfer,
S Desai,
K Georgatzis,
P Beaumont, and
B Aragam

AISTATS 2020

proceedings /
preprint

We revisit the structure learning problem for dynamic Bayesian networks and propose a method that simultaneously estimates contemporaneous (intra-slice) and time-lagged (inter-slice) relationships between variables in a time-series. Our approach is score-based, and revolves around minimizing a penalized loss subject to an acyclicity constraint. To solve this problem, we leverage a recent algebraic result characterizing the acyclicity constraint as a smooth equality constraint. The resulting algorithm, which we call DYNOTEARS, outperforms other methods on simulated data, especially in high-dimensions as the number of variables increases. We also apply this algorithm on real datasets from two different domains, finance and molecular biology, and analyze the resulting output. Compared to state-of-the-art methods for learning dynamic Bayesian networks, our method is both scalable and accurate on real data. The simple formulation, competitive performance, and scalability of our method make it suitable for a variety of problems where one seeks to learn connections between variables across time.

Keywords: time series, dynamic Bayesian networks, continuous optimization, graphical models, structural vector autoregression

Learning sample-specific models with low-rank personalized regression

B Lengerich, B Aragam, and EP Xing

NeurIPS 2019

proceedings /
preprint /
code

Modern applications of machine learning (ML) deal with increasingly heterogeneous datasets comprised of data collected from overlapping latent subpopulations. As a result, traditional models trained over large datasets may fail to recognize highly predictive localized effects in favour of weakly predictive global patterns. This is a problem because localized effects are critical to developing individualized policies and treatment plans in applications ranging from precision medicine to advertising. To address this challenge, we propose to estimate sample-specific models that tailor inference and prediction at the individual level. In contrast to classical ML models that estimate a single, complex model (or only a few complex models), our approach produces a model personalized to each sample. These sample-specific models can be studied to understand subgroup dynamics that go beyond coarse-grained class labels. Crucially, our approach does not assume that relationships between samples (e.g. a similarity network) are known a priori. Instead, we use unmodeled covariates to learn a latent distance metric over the samples. We apply this approach to financial, biomedical, and electoral data as well as simulated data and show that sample-specific models provide fine-grained interpretations of complicated phenomena without sacrificing predictive accuracy compared to state-of-the-art models such as deep neural networks.

Keywords: personalization, sample-specific, low-rank models, personalized regression

Globally optimal score-based learning of directed acyclic graphs in high-dimensions

B Aragam, AA Amini, and Q Zhou

NeurIPS 2019

proceedings /
preprint

We prove that $\Omega(s\log p)$ samples suffice to learn a sparse Gaussian directed acyclic graph (DAG) from data, where $s$ is the maximum Markov blanket size. This improves upon recent results that require $\Omega(s^{4}\log p)$ samples in the equal variance case. To prove this, we analyze a popular score-based estimator that has been the subject of extensive empirical inquiry in recent years and is known to achieve state-of-the-art results. Furthermore, the approach we study does not require strong assumptions such as faithfulness that existing theory for score-based learning crucially relies on. The resulting estimator is based around a difficult nonconvex optimization problem, and its analysis may be of independent interest given recent interest in nonconvex optimization in machine learning. Our analysis overcomes the drawbacks of existing theoretical analyses, which either fail to guarantee structure consistency in high-dimensions (i.e. learning the correct graph with high probability), or rely on restrictive assumptions. In contrast, we give explicit finite-sample bounds that are valid in the important $p\gg n$ regime.

Keywords: graphical modeling, directed acyclic graphs, sample complexity, score-based learning

Identifiability of nonparametric mixture models and Bayes optimal clustering

B Aragam, C Dan, EP Xing, and P Ravikumar

Annals of Statistics

journal /
preprint

Motivated by problems in data clustering, we establish general conditions under which families of nonparametric mixture models are identifiable by introducing a novel framework for clustering overfitted parametric (i.e. misspecified) mixture models. These conditions generalize existing conditions in the literature, and are flexible enough to include for example mixtures of Gaussian mixtures. In contrast to the recent literature on estimating nonparametric mixtures, we allow for general nonparametric mixture components, and instead impose regularity assumptions on the underlying mixing measure. As our primary application, we apply these results to partition-based clustering, generalizing the well-known notion of a Bayes optimal partition from classical model-based clustering to nonparametric settings. Furthermore, this framework is constructive in that it yields a practical algorithm for learning identified mixtures, which is illustrated through several examples. The key conceptual device in the analysis is the convex, metric geometry of probability distributions on metric spaces and its connection to optimal transport and the Wasserstein convergence of mixing measures. The result is a flexible framework for nonparametric clustering with formal consistency guarantees.

Keywords: nonparametric statistics, mixture models, clustering, identifiability, optimal transport

Fault tolerance in iterative-convergent machine learning

A Qiao, B Aragam, B Zhang, and EP Xing

ICML 2019

proceedings /
preprint

Machine learning (ML) training algorithms often possess an inherent self-correcting behavior due to their iterative-convergent nature. Recent systems exploit this property to achieve adaptability and efficiency in unreliable computing environments by relaxing the consistency of execution and allowing calculation errors to be self-corrected during training. However, the behavior of such systems are only well understood for specific types of calculation errors, such as those caused by staleness, reduced precision, or asynchronicity, and for specific types of training algorithms, such as stochastic gradient descent. In this paper, we develop a general framework to quantify the effects of calculation errors on iterative-convergent algorithms and use this framework to design new strategies for checkpoint-based fault tolerance. Our framework yields a worst-case upper bound on the iteration cost of arbitrary perturbations to model parameters during training. Our system, SCAR, employs strategies which reduce the iteration cost upper bound due to perturbations incurred when recovering from checkpoints. We show that SCAR can reduce the iteration cost of partial failures by 78%--95% when compared with traditional checkpoint-based fault tolerance across a variety of ML models and training algorithms.

Keywords: fault tolerance, distributed systems, machine learning, reliability, iterative algorithms

The sample complexity of semi-supervised learning with nonparametric mixture models

C Dan, L Leqi, B Aragam, P Ravikumar, and EP Xing

NeurIPS 2018

proceedings /
preprint

We study the sample complexity of semi-supervised learning (SSL) and introduce new assumptions based on the mismatch between a mixture model learned from unlabeled data and the true mixture model induced by the (unknown) class conditional distributions. Under these assumptions, we establish an \Omega(K\log K) labeled sample complexity bound without imposing parametric assumptions, where K is the number of classes. Our results suggest that even in nonparametric settings it is possible to learn a near-optimal classifier using only a few labeled samples. Unlike previous theoretical work which focuses on binary classification, we consider general multiclass classification (K>2), which requires solving a difficult permutation learning problem. This permutation defines a classifier whose classification error is controlled by the Wasserstein distance between mixing measures, and we provide finite-sample results characterizing the behaviour of the excess risk of this classifier. Finally, we describe three algorithms for computing these estimators based on a connection to bipartite graph matching, and perform experiments to illustrate the superiority of the MLE over the majority vote estimator.

Keywords: semi-supervised learning, mixture models, nonparametric statistics, permutation learning, coupon collection, sample complexity

DAGs with NO TEARS: Continuous optimization for structure learning

X Zheng, B Aragam, P Ravikumar, and EP Xing

NeurIPS 2018 (spotlight)

proceedings /
preprint /
code /
blog

Estimating the structure of directed acyclic graphs (DAGs, also known as Bayesian networks) is a challenging problem since the search space of DAGs is combinatorial and scales superexponentially with the number of nodes. Existing approaches rely on various local heuristics for enforcing the acyclicity constraint. In this paper, we introduce a fundamentally different strategy: we formulate the structure learning problem as a purely continuous optimization problem over real matrices that avoids this combinatorial constraint entirely. This is achieved by a novel characterization of acyclicity that is not only smooth but also exact. The resulting problem can be efficiently solved by standard numerical algorithms, which also makes implementation effortless. The proposed method outperforms existing ones, without imposing any structural assumptions on the graph such as bounded treewidth or in-degree.

Keywords: directed acyclic graphs, Bayesian networks, constrained optimization, nonconvex optimization, augmented Lagrangian, black-box

Personalized regression enables sample-specific pan-cancer analysis

B Lengerich, B Aragam, and EP Xing

Bioinformatics (also appears in ISMB 2018)

journal / preprint / code

In many applications, inter-sample heterogeneity is crucial to understanding the complex biological processes under study. For example, in genomic analysis of cancers, each patient in a cohort may have a different driver mutation, making it difficult or impossible to identify causal mutations from an averaged view of the entire cohort. Unfortunately, many traditional methods for genomic analysis seek to estimate a single model which is shared by all samples in a population, ignoring this inter-sample heterogeneity entirely. In order to better understand patient heterogeneity, it is necessary to develop practical, personalized statistical models. To uncover this inter-sample heterogeneity, we propose a novel regularizer for achieving patient-specific personalized estimation. This regularizer operates by learning latent distance metrics between personalized parameters and clinical covariates, and attempting to match these distances as closely as possible. Crucially, we do not assume these distances are already known. Instead, we allow the data to dictate the structure of these latent distance metrics. Finally, we apply our method to learn patient-specific, interpretable models for a pan-cancer gene expression dataset containing samples from more than 30 distinct cancer types and find strong evidence of personalization effects between cancer types as well as between individuals. Our analysis uncovers sample-specific aberrations that are overlooked by population-level methods, suggesting a promising new path for precision analysis of complex diseases such as cancer.

Keywords: precision medicine, personalized regression, patient-specific modeling, distance-matching, TCGA

Precision lasso: Accounting for correlations and linear dependencies in high-dimensional genomic data

H Wang, B Lengerich, B Aragam, and EP Xing

Bioinformatics

journal / code

Keywords: high-dimensional regression, genomics, irrepresentability, correlated variables

Variable selection in heterogeneous datasets: A truncated-rank sparse linear mixed model with applications to genome-wide association studies

H Wang, B Aragam, and EP Xing

Methods

journal / preprint

A fundamental and important challenge in modern datasets of ever increasing dimensionality is variable selection, which has taken on renewed interest recently due to the growth of biological and medical datasets with complex, non-i.i.d. structures. Naively applying classical variable selection methods such as the Lasso to such datasets may lead to a large number of false discoveries. Motivated by genome-wide association studies in genetics, we study the problem of variable selection for datasets arising from multiple subpopulations, when this underlying population structure is unknown to the researcher. We propose a unified framework for sparse variable selection that adaptively corrects for population structure via a low-rank linear mixed model. Most importantly, the proposed method does not require prior knowledge of sample structure in the data and adaptively selects a covariance structure of the correct complexity. Through extensive experiments, we illustrate the effectiveness of this framework over existing methods. Further, we test our method on three different genomic datasets from plants, mice, and human, and discuss the knowledge we discover with our method.

Keywords: GWAS, linear mixed models, heterogeneous data, confounding, population structure

Learning directed acyclic graphs with penalized neighbourhood regression

B Aragam, AA Amini, and Q Zhou

preprint

We study a family of regularized score-based estimators for learning the structure of a directed acyclic graph (DAG) for a multivariate normal distribution from high-dimensional data with p >> n. Our main results establish support recovery guarantees and deviation bounds for a family of penalized least-squares estimators under concave regularization without assuming prior knowledge of a variable ordering. These results apply to a variety of practical situations that allow for arbitrary nondegenerate covariance structures as well as many popular regularizers including the MCP, SCAD, L0 and L1. The proof relies on interpreting a DAG as a recursive linear structural equation model, which reduces the estimation problem to a series of neighbourhood regressions. We provide a novel statistical analysis of these neighbourhood problems, establishing uniform control over the superexponential family of neighbourhoods associated with a Gaussian distribution. We then apply these results to study the statistical properties of score-based DAG estimators, learning causal DAGs, and inferring conditional independence relations via graphical models. Our results yield---for the first time---finite-sample guarantees for structure learning of Gaussian DAGs in high-dimensions via score-based estimation.

Keywords: graphical modeling, high-dimensional statistics, concave regularization, directed acyclic graphs, structural equations, sparse regression

Learning large-scale Bayesian networks with the sparsebn package

B Aragam, J Gu, and Q Zhou

Journal of Statistical Software

journal /
preprint /
code

Learning graphical models from data is an important problem with wide applications, ranging from genomics to the social sciences. Nowadays datasets often have upwards of thousands---sometimes tens or hundreds of thousands---of variables and far fewer samples. To meet this challenge, we have developed a new R package called sparsebn for learning the structure of large, sparse graphical models with a focus on Bayesian networks. While there are many existing software packages for this task, this package focuses on the unique setting of learning large networks from high-dimensional data, possibly with interventions. As such, the methods provided place a premium on scalability and consistency in a high-dimensional setting. Furthermore, in the presence of interventions, the methods implemented here achieve the goal of learning a causal network from data. Additionally, the sparsebn package is fully compatible with existing software packages for network analysis.

Keywords: R, software, graphical modeling, directed acyclic graphs, structural equations

Concave penalized estimation of sparse Gaussian Bayesian networks

B Aragam and Q Zhou

Journal of Machine Learning Research

journal /
preprint /
code

We develop a penalized likelihood estimation framework to learn the structure of Gaussian Bayesian networks from observational data. In contrast to recent methods which accelerate the learning problem by restricting the search space, our main contribution is a fast algorithm for score-based structure learning which does not restrict the search space in any way and works on high-dimensional data sets with thousands of variables. Our use of concave regularization, as opposed to the more popular L0 (e.g. BIC) penalty, is new. Moreover, we provide theoretical guarantees which generalize existing asymptotic results when the underlying distribution is Gaussian. Most notably, our framework does not require the existence of a so-called faithful DAG representation, and as a result, the theory must handle the inherent nonidentifiability of the estimation problem in a novel way. Finally, as a matter of independent interest, we provide a comprehensive comparison of our approach to several standard structure learning methods using open-source packages developed for the R language. Based on these experiments, we show that our algorithm obtains higher sensitivity with comparable false discovery rates for high-dimensional data and scales efficiently as the number of nodes increases. In particular, the total runtime for our method to generate a solution path of 20 estimates for DAGs with 8000 nodes is around one hour.

Keywords: Bayesian networks, concave penalization, directed acyclic graphs, coordinate descent, nonconvex optimization

// theses

Research into graphical models is a rapidly developing enterprise, garnering significant interest from both the statistics and machine learning communities. A parallel thread in both communities has been the study of low-dimensional structures in high-dimensional models where $p\gg n$. Recently, there has been a surge of interest in connecting these threads in order to understand the behaviour of graphical models in high-dimensions. Due to their relative simplicity, undirected models such as the Gaussian graphical model and Ising models have received most of the attention, whereas directed graphical models have received comparatively little attention. An important yet largely unresolved class of directed graphical models are Bayesian networks, or directed acyclic graphs (DAGs). These models have a wide variety of applications in aritificial intelligence, machine learning, genetics, and computer vision, but estimation of Bayesian networks in high-dimensions is not well-understood. The main focus of this dissertation is to address some fundamental questions about these models in high-dimensions.

The primary goal is to develop both algorithms and theory for estimating continuous, linear Bayesian networks, capable of handling modern high-dimensional problems. Motivated by problems from the regression literature, we show how to adapt recent work in sparse learning and nonconvex optimization to the structure learning problem for Bayesian networks in order to estimate DAGs with several thousand nodes. We draw an explicit connection between linear Bayesian networks and so-called neighbourhood regression problems and show how this can be exploited in order to derive nonasymptotic performance bounds for penalized least squares estimators of directed graphical models.

On the algorithmic side, we develop a method for estimating Gaussian Bayesian networks based on convex reparametrization and cyclic coordinate descent. In contrast to recent methods which accelerate the learning problem by restricting the search space, we propose a method for score-based structure learning which does not restrict the search space. We do not require the existence of a so-called faithful DAG representation, and as a result, our methodology must handle the inherent nonidentifiability of the estimation problem in a novel way. On the theoretical side, we provide (a) Finite-dimensional performance guarantees for local minima of the resulting nonconvex program, and (b) A general high-dimensional framework for global minima of the nonconvex program. Both the algorithms and theory apply to a general class of regularizers, including the MCP, SCAD, $\ell_1$ and $\ell_0$ penalties. Finally, as a matter of independent interest, we provide a comprehensive comparison of our approach to several standard structure learning methods using open-source packages developed for the \texttt{R} language.

Keywords: Bayesian networks, high-dimensional statistics, graphical models, sparse regression, concave regularization, nonconvex optimization

Volume comparison with integral bounds in Lorentz manifolds

Undergraduate Thesis

Ten years ago, Ehrlich and Sanchez produced a pointwise statement of the classical Bishop volume comparison theorem for so-called SCLV subsets of the causal future in a Lorentz manifold, while Petersen and Wei developed and proved an integral version for Riemannian manifolds. We apply Peterson and Wei's method to the SCLV sets, and verify that two essential differential equations from the Riemannian proof extend to the Lorentz setting. As a result, we obtain a volume comparison theorem for Lorentz manifolds with integral, rather than pointwise, bounds. We also briey discuss the history of the problem, starting with Bishop's original theorem from 1963.

Keywords: differential geometry, volume comparison, Lorentz manifolds

// current

Wai Ming Tai (postdoc)

Kevin Bello (postdoc, joint with Pradeep Ravikumar)

Chang Deng (PhD student, Booth)

Yibo Jiang (PhD student, CS / DSI, joint with Victor Veitch)

Zhao Lyu (PhD student, Applied Math / Statistics)

Mingyu Liu (MS student, Statistics)

// past

Yi Ding (CS PhD @ UChicago -> postdoc @ MIT)

Ruiyi Yang (Applied Math PhD @ UChicago -> postdoc @ Princeton)

Bohdan Kivva (CS/Math PhD @ UChicago -> Google)

Goutham Rajendran (CS PhD @ UChicago -> postdoc @ Carnegie Mellon University)

Karhan Kayan (CS/Math undergrad @ UChicago -> CS PhD @ Princeton)

I will not be teaching in Fall 2019 or Winter 2020.

Past teaching assignments:

- Machine Learning 10-821: Data Analysis Project Preparation (Fall 2017)
- Statistics 10: Introduction to Statistical Reasoning (Spring 2016)
- Statistics 10: Introduction to Statistical Reasoning (Winter 2016)
- Statistics 10: Introduction to Statistical Reasoning (Fall 2015)
- Statistics 495A: Teaching College Statistics (Winter 2015)
- Statistics 100A: Introduction to Probability (Spring 2014)
- Statistics 101B: Introduction to Design and Analysis of Experiments (Winter 2014)
- Statistics 102A: Introduction to Computational Statistics with R (Fall 2013)
- PIC 20A: Principles of Java (Spring 2010)
- PIC 10A: Introduction to C++ Programming (Winter 2010)
- PIC 10A: Introduction to C++ Programming (Fall 2009)

// software

You can find more up-to-date information on my software projects by visiting my Github page.

// DAGMA

Faster and more accurate continuous constrained optimization for structure learning based on a new acyclicity characterization via the log-det function. Can be used with any loss function and includes implementations for both linear and nonlinear (e.g. neural network) models.

// TAM: Learning DAGs with testing and masking

The TAM algorithm is used for learning the structure of a directed acyclic graph (DAG). Given data from a nonparametric distribution that satisfies an entropic condition, TAM efficiently learns the DAG that generated the samples.

// Automated dependence plots

Python library for auditing, checking, and explaining black-box machine learning models by automating the selection of interesting dependence plots. Highlights surprising or undesirable model behaviours as linear combinations of raw features or as paths in a latent space arising from a generative model.

// NO TEARS

DAG learning formulated as a continuous, black-box optimization problem over real matrices that avoids combinatorial optimization. This repository includes two versions: A simple version that is implemented using scipy in fewer than 50 lines of Python code, and an L1-regularized version for both linear and nonlinear models.

// Precision Lasso

The Precision Lasso is a variant of the Lasso designed to adapt to and account for correlations and dependencies in high-dimensional data.

// Personalized regression

This repository contains Python code for learning sample-specific, personalized regression models. The goal of personalized regression is to perform retrospective analysis by estimating simple models that each apply to a single sample. After estimating these sample-specific models, we have a matrix of model parameters which we may analyze as we wish.

// sparsebn package for R

sparsebn is an R package for learning large-scale Bayesian networks from high-dimensional data. It allows users to incorporate mixed experimental and observational data with either continuous or discrete observations, and scales to datasets with many thousands of variables. The underlying framework is based on recent developments in sparse (e.g. L1) regularization, coordinate descent, and nonconvex optimization.

// ccdr package for R

The source code for the CCDr algorithm described in Aragam and Zhou (2015) is freely available online through GitHub.

ccdr is an R package for structure learning of linear Bayesian networks from high-dimensional, Gaussian data. The underlying algorithm estimates a Bayesian network (aka DAG or belief net) using penalized maximum likelihood based on L1 or concave (MCP) regularization and observational data.

// contact

// contact

Office: Harper Center 446

Email: bryon at chicagobooth dot edu

Phone: 773-834-5892

Booth School of Business

5807 S Woodlawn Ave

Chicago, IL 60637

I also got a little bored while designing this site so I hid some easter eggs here and there.

// biography

Bryon Aragam studies statistical machine learning, nonparametric statistics, and unsupervised learning. His current interests involve (i) Statistical aspects of latent variable models, (ii) Model selection and identification in nonparametric models, and (iii) Theory and computation for deep generative models. In particular, this work focuses on applications of artificial intelligence, including tools such as ChatGPT and DALL-E. His work attempts to understand the statistical foundations of these models and how to improve them from both practical and theoretical perspectives. He is also involved with developing open-source software and solving problems in interpretability, ethics, and fairness in artificial intelligence.

Prior to joining the University of Chicago, he was a project scientist and postdoctoral researcher in the Machine Learning Department at Carnegie Mellon University. He completed his PhD in Statistics and a Masters in Applied Mathematics at UCLA, where he was an NSF graduate research fellow. Bryon has also served as a data science consultant for technology and marketing firms, where he has worked on problems in survey design and methodology, ranking, customer retention, and logistics.