Papers
Topics
Authors
Recent
Search
2000 character limit reached

Chow–Liu Tree Models

Updated 5 February 2026
  • Chow–Liu tree is a graphical model that approximates high-dimensional joint distributions by factoring them into tree-structured conditional probabilities using mutual information.
  • It computes pairwise mutual information and employs efficient algorithms like Kruskal’s or Prim’s to construct an optimal tree-structured model.
  • Extensions include causal, conditional, and forest variants, enabling applications in network inference, data compression, and probabilistic modeling.

A Chow–Liu tree is a graphical model that provides an optimal tree-structured approximation to a high-dimensional joint probability distribution by leveraging pairwise mutual informations as edge weights. Originally formulated for discrete variables, the concept generalizes to continuous (e.g., Gaussian), mixed, temporal, and conditional settings. The central algorithmic idea is to reduce maximum-likelihood estimation over tree-structured models to a maximum-weight spanning tree computation, with theoretical and computational guarantees that distinguish it from more complex graphical model learning problems.

1. Foundational Principles and Objective Function

Given mm random variables X1,...,XmX_1, ..., X_m (discrete, continuous, or mixed), the Chow–Liu algorithm seeks a tree-structured model P^T\widehat{P}_T that minimizes the Kullback–Leibler (KL) divergence to the true joint law P(X1,...,Xm)P(X_1,...,X_m): P^T=argminTD(PP^T)\widehat{P}_T = \arg\min_{T} D\bigl(P \,\|\, \widehat{P}_T\bigr) where P^T\widehat{P}_T factorizes over the edges (i,j)T(i,j)\in T of a tree: P^T(x1,,xm)=(i,j)TP(xixj)\widehat{P}_T(x_1,\ldots,x_m) = \prod_{(i,j)\in T} P(x_i|x_j) Chow and Liu [1968] showed that this minimization is equivalent to maximizing the sum of pairwise mutual informations over the tree: argminTD(PP^T) = argmaxT(i,j)TI(Xi;Xj)\arg\min_T D(P\,\|\,\widehat{P}_T)\ =\ \arg\max_T \sum_{(i,j)\in T} I(X_i; X_j) where

I(Xi;Xj)=xi,xjPij(xi,xj)logPij(xi,xj)Pi(xi)Pj(xj)I(X_i;X_j)=\sum_{x_i,x_j} P_{ij}(x_i,x_j)\log \frac{P_{ij}(x_i,x_j)}{P_i(x_i)P_j(x_j)}

This result applies for both discrete and multivariate Gaussian distributions, though in the Gaussian case the mutual information reduces to I(Xi;Xj)=12log(1ρij2)I(X_i;X_j)=-\frac{1}{2}\log(1-\rho_{ij}^2), where ρij\rho_{ij} is the correlation coefficient.

2. Algorithmic Methodology

The standard Chow–Liu procedure consists of the following steps:

  1. Compute Mutual Information Weights: For each unordered pair (i,j)(i,j), estimate I(Xi;Xj)I(X_i; X_j) using empirical frequencies (for discrete) or empirical covariances and correlations (for Gaussian).
  2. Build Complete Weighted Graph: Assign weight wij=I(Xi;Xj)w_{ij} = I(X_i; X_j) to each edge.
  3. Maximum-Weight Spanning Tree (MWST): Use Kruskal’s or Prim’s algorithm to find the spanning tree TT^* maximizing the total weight (i,j)Twij\sum_{(i,j)\in T} w_{ij}.
  4. Model Construction: The ML (maximum-likelihood) parameters for the selected tree are the empirical marginals on the edges: QT(xi,xj)=P^(xi,xj)Q_T(x_i,x_j)=\widehat{P}(x_i,x_j). The complete factorization uses either parent conditionals or pairwise marginals over the spanning tree.

Computational Complexity: Calculating all pairwise mutual informations and running MWST is O(m2n)O(m^2 n) (empirical MI estimation) plus O(m2logm)O(m^2 \log m) (MWST) for mm variables (Srebro, 2013). For large mm, the method remains tractable and scales well in practical high-dimensional settings (Tan et al., 2010, Wang et al., 2024).

3. Extensions: Polytrees, Causal Trees, Forests, and Conditional Structures

Polytrees

A polytree is a directed acyclic graph (DAG) whose underlying undirected graph is a tree. Branchings (directed trees) can be found using Chow–Liu, but optimal polytree learning is NP-hard even for degree-2 polytrees. The Chow–Liu branching provides a provable logarithmic approximation to the ML polytree. Let U=maxiH(Xi), L=miniH(Xi)U = \max_i H(X_i),\ L = \min_i H(X_i), then for the Chow–Liu tree TT^* and ML polytree GG^*: Cost(T)αCost(G)whereα=7/2+12log(U/L)\text{Cost}(T^*) \leq \alpha \cdot \text{Cost}(G^*)\quad \text{where}\quad \alpha=7/2+ \frac{1}{2}\log(U/L) No polynomial-time algorithm can achieve a strictly better constant-factor on general data (Dasgupta, 2013).

Causal Chow–Liu Trees

For time-series or multivariate processes, the classical Chow–Liu tree does not respect temporal causality. A causal version replaces mutual information with directed information: I(XnYn)=t=1nI(Xt;YtYt1)I(X^n \to Y^n) = \sum_{t=1}^n I(X^t;Y_t|Y^{t-1}) Causal Chow–Liu trees maximize the sum of directed informations over a directed spanning arborescence, solvable efficiently via the Chu–Liu/Edmonds algorithm. This construction preserves temporal and causal orderings and is suitable for modeling dynamical systems (Quinn et al., 2011).

Forests and MDL Penalization

Unpenalized Chow–Liu trees may overfit when the true structure is a forest (union of trees). Pruning by thresholding mutual informations (CLThres) enables structural and risk consistency when the true distribution is forest-structured (Tan et al., 2010). The minimum description length (MDL) approach penalizes edge additions by complexity, yielding model and parameter selection for mixed-discrete/Gaussian data and supports learning generalized forests (Suzuki, 2010).

Conditional and Emission Chow–Liu Trees

Conditional Chow–Liu trees extend the method to conditional densities, producing optimal tree-structured factorizations P(XZ)P(X|Z) where statistical dependencies among XX are conditional on ZZ. In hidden Markov models, tree-structured conditional emission distributions (HMM-CL, HMM-CCL) provide parsimonious and interpretable models for vector time-series and demonstrate state-of-the-art empirical performance for high-dimensional sequence data (Kirshner et al., 2012).

4. Statistical Guarantees and Sample Complexity

Exact and Approximate Structure Recovery

For dd variables over alphabet size nn and exact recovery in the noiseless case, the sufficient number of samples is governed by the smallest mutual information gap τ\tau ("information threshold") between true edges and best non-edges: n=O(log(d/δ)τ2)n = O\left(\frac{\log(d/\delta)}{\tau^2}\right) No algorithm (including Chow–Liu) can succeed with fewer samples up to constant factors (Nikolakakis et al., 2019, 0905.0940, Bhattacharyya et al., 2020).

Learning in KL, Total Variation, and Local TV Loss

  • Proper learning (minimizing DKL(PQT)D_{\mathrm{KL}}(P \| Q_T)) in the realizable-tree case is achievable at N=O~(Σ3n/ε)N=\widetilde{O}(|\Sigma|^3 n/\varepsilon) samples for accuracy ε\varepsilon in KL (Bhattacharyya et al., 2020).
  • For tree-structured Ising models, proper learning in total variation distance ϵ\epsilon is sample-optimal at O(nlnn/ϵ2)O(n \ln n / \epsilon^2) (Daskalakis et al., 2020).
  • Under prediction-centric local total variation, Chow–Liu achieves optimal rates on tree-Ising distributions only when edge strengths are bounded; Chow–Liu++ achieves the information-theoretic optimal O(logn/ϵ2)O(\log n/\epsilon^2) rate robustly (Boix-Adsera et al., 2021).

Noisy or Hidden Models

When data are corrupted by known or unknown noise, the minimal sample size for correct structure recovery remains governed by the post-noise information threshold. Preprocessing (e.g., channel whitening) may be necessary to maintain identifiability if noise can cause threshold collapse (Nikolakakis et al., 2019).

5. Connections to Bayesian Inference and Model Selection

The Chow–Liu maximum-weight spanning tree provides the mode (MAP) of the posterior when trees have edge-factorizable priors. Using the Matrix Tree Theorem, it is possible to average quantities over the full posterior distribution on trees in O(d3)O(d^3) time and thus perform Bayesian model averaging efficiently in the tree-structured class. The approach generalizes to forests, polytree models, and mixed variable types with appropriate priors and computational routines (Jones, 2021).

6. Applications and Impact

Chow–Liu trees underpin graphical model selection, density estimation, structure discovery in biological and social networks, and are routinely used as submodules in latent variable models, hierarchical learning, and generative models based on tree tensor networks. Tree-based representations have been employed for efficient compression, denoising, forecasting, and probabilistic inference across statistics, machine learning, and signal processing, particularly where interpretability and computational speed are required at scale (Tan et al., 2010, Tang et al., 2022).

7. Limitations and Open Directions

Chow–Liu trees inherit the expressiveness limitation of tree factorizations: in domains with loops or higher-order dependencies, their approximation error may be significant. Extension to bounded tree-width Markov networks is NP-hard; learning polytrees is also NP-hard to approximate beyond a fixed constant. Chow–Liu is globally optimal for tree-structured approximations but not robust to certain model misspecifications. Recent advances such as Chow–Liu++ address prediction-centric objectives, distributional robustness, and adversarial contamination (Boix-Adsera et al., 2021).

Open directions include efficient learning of higher-treewidth models, generalization to mixed or complex data types, and extending locally optimal learning guarantees beyond the tree class (Dasgupta, 2013, Boix-Adsera et al., 2021, Wang et al., 2024).


References

  • "Causal Dependence Tree Approximations of Joint Distributions for Multiple Random Processes" (Quinn et al., 2011)
  • "Learning Polytrees" (Dasgupta, 2013)
  • "A Large-Deviation Analysis of the Maximum-Likelihood Learning of Markov Tree Structures" (0905.0940)
  • "Learning High-Dimensional Markov Forest Distributions: Analysis of Error Rates" (Tan et al., 2010)
  • "Optimal estimation of Gaussian (poly)trees" (Wang et al., 2024)
  • "Sample-Optimal and Efficient Learning of Tree Ising models" (Daskalakis et al., 2020)
  • "A Generalization of the Chow-Liu Algorithm and its Application to Statistical Learning" (Suzuki, 2010)
  • "Bayesian learning of forest and tree graphical models" (Jones, 2021)
  • "Maximum Likelihood Bounded Tree-Width Markov Networks" (Srebro, 2013)
  • "Conditional Chow-Liu Tree Structures for Modeling Discrete-Valued Vector Time Series" (Kirshner et al., 2012)
  • "Chow-Liu++: Optimal Prediction-Centric Learning of Tree Ising Models" (Boix-Adsera et al., 2021)
  • "Optimal Rates for Learning Hidden Tree Structures" (Nikolakakis et al., 2019)
  • "Near-Optimal Learning of Tree-Structured Distributions by Chow-Liu" (Bhattacharyya et al., 2020)
  • "Generative Modeling via Tree Tensor Network States" (Tang et al., 2022)
  • "Latent Tree Approximation in Linear Model" (Khajavi, 2017)
  • "Decentralized Learning of Tree-Structured Gaussian Graphical Models from Noisy Data" (Hussain, 2021)
  • "An Entropy-based Learning Algorithm of Bayesian Conditional Trees" (Geiger, 2013)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Chow–Liu Tree.