Papers
Topics
Authors
Recent
Search
2000 character limit reached

Composite Quasi-Likelihood (CQL)

Updated 20 February 2026
  • Composite Quasi-Likelihood (CQL) is a method for estimating group-specific dynamic panel models with spatial correlation while addressing fixed effect misspecification.
  • It leverages pairwise Gaussian quasi-likelihoods and the DC algorithm to jointly optimize model parameters and latent group classifications in high-dimensional settings.
  • Empirical results show high classification accuracy (>95%) and reduced bias, validating its use in complex spatial dynamic panel models.

Composite quasi-likelihood (CQL) refers to a quasi-likelihood methodology for estimation and classification in large dynamic panel models with group-specific heterogeneity and spatially dependent errors. Developed to address the computational infeasibility of the full likelihood in high-dimensional panels with unobserved group structure and spatial correlation, the CQL approach leverages composite (instrumental or marginal) likelihoods, paired with sophisticated optimization algorithms, to enable consistent and efficient parameter and group membership estimation in models subject to fixed effects misspecification and spatial dependence (Chu, 2017).

1. Model Framework and Specification

The foundational setting is a balanced panel {yit,xit}\{y_{it}, x_{it}\}, i=1,,Ni=1, \dots, N, t=1,,Tt=1, \dots, T. Each observational unit ii belongs to one of GG unknown latent groups, gi{1,,G}g_i \in \{1, \dots, G\}. Conditional on group membership, the core dynamic error-correction model is: yit=αi+βgixit+ϕgi(yi,t1θgixi,t1)+uity_{it} = \alpha_i + \beta_{g_i}' x_{it} + \phi_{g_i}\left(y_{i,t-1} - \theta_{g_i}' x_{i,t-1} \right) + u_{it} where:

  • αi\alpha_i is an individual fixed effect (which may be misspecified),
  • (βg,ϕg,θg)(\beta_g, \phi_g, \theta_g) are group-specific slope and adjustment parameters,
  • uitu_{it} is an innovation permitting both serial and (contemporaneous) spatial correlation.

Spatial error dependence is modeled as

ut=Wut+εtu_t = W u_t + \varepsilon_t

where ut=(u1t,,uNt)u_t = (u_{1t}, \dots, u_{Nt})', WW is a spatial-weights matrix (diagonal elements zero), and εt\varepsilon_t is spherical, E[εtεt]=σ2INE[\varepsilon_t \varepsilon_t'] = \sigma^2 I_N. This yields Cov(ut)=σ2(IW)1(IW)1\text{Cov}(u_t) = \sigma^2 (I-W)^{-1}(I-W')^{-1} (Chu, 2017).

2. Composite Quasi-Likelihood Objective

The infeasibility of the NN-dimensional Gaussian likelihood at large NN motivates the use of a composite likelihood. Specifically, the CQL objective is formed by summing pairwise (bivariate) Gaussian quasi-log-likelihood contributions: LCQL(ψ,{gi}i=1N)=i<jwijt=2Tij,t(ψ;(yit,yjt),(yi,t1,yj,t1),(xit,xjt),(xi,t1,xj,t1))\mathcal{L}_{CQL}(\psi, \{g_i\}_{i=1}^N) = \sum_{i<j} w_{ij} \sum_{t=2}^T \ell_{ij, t}\left(\psi; (y_{it}, y_{jt}), (y_{i, t-1}, y_{j, t-1}), (x_{it}, x_{jt}), (x_{i, t-1}, x_{j, t-1})\right) where

  • ψ\psi is the collection of all structural parameters,
  • wijw_{ij} are user-specified weights (commonly wij=1w_{ij}=1),
  • ij,t\ell_{ij, t} is the Gaussian quasi-log-likelihood for (i,j)(i, j), constructed using the 2×2 implied covariance from the spatial process (Chu, 2017).

3. Joint Parameter Estimation and Latent Group Classification

Both group memberships {gi}\{g_i\} and model parameters ψ\psi are estimated jointly by maximizing the composite quasi-likelihood. This leads to the optimization problem: minψ,{gi}LCQL(ψ,{gi})s.t. gi{1,,G}, g=1G1{gi=g}=1 i\min_{\psi, \{g_i\}} -\mathcal{L}_{CQL}(\psi, \{g_i\}) \quad \text{s.t.} \ g_i \in \{1, \dots, G\},\ \sum_{g=1}^G 1\{g_i=g\} = 1 \ \forall i Introducing binary indicators zi,g=1{gi=g}z_{i,g} = 1\{g_i=g\}, the problem becomes a nonconvex mixed-integer program (MIP), as the group assignments interact nonlinearly with both the likelihood and spatial covariance structure (Chu, 2017).

4. Optimization via Difference-of-Convex Decomposition and DCA

Chu demonstrates that LCQL-\mathcal{L}_{CQL}, given fixed Z={zi,g}Z = \{z_{i,g}\}, can be decomposed as the difference of two proper convex functions in the continuous parameters: Q(ψ,Z)=f(ψ,Z)g(ψ,Z)Q(\psi, Z) = f(\psi, Z) - g(\psi, Z) where ff captures convexified quadratic forms (log-determinant and trace terms, regularized as needed) and gg corrects over-regularization to restore the exact composite objective. This structure enables the application of the DC Algorithm (DCA):

  • At each iteration with membership fixed, compute a subgradient of gg in ψ\psi.
  • Minimize the resulting strongly convex surrogate to update ψ\psi.
  • Optionally, update group memberships ZZ by re-assigning each ii to the group that minimizes its local objective, or perform a discrete optimization step.
  • Alternate until convergence.

This iterative block-coordinate approach efficiently heuristically solves the MIP, providing a scalable solution route for high-dimensional panels (Chu, 2017).

5. Theoretical Properties

Under moderate regularity:

  • Asymptotic unbiasedness and consistency of the slope parameters (βg,ϕg,θg)(\beta_g, \phi_g, \theta_g) holds jointly as N,TN, T \to \infty, even when individual fixed effects αi\alpha_i are arbitrary or misspecified.
  • The estimator attains the oracle property: estimation and classification are such that the limiting distribution of the parameters matches that with perfect (oracle) knowledge of group memberships and fixed effects.
  • Asymptotic normality holds for each group: NgT(β^gβg0 ϕ^gϕg0 θ^gθg0)N(0,Vg)\sqrt{N_g T} \begin{pmatrix} \widehat{\beta}_g - \beta_g^0 \ \widehat{\phi}_g - \phi_g^0 \ \widehat{\theta}_g - \theta_g^0 \end{pmatrix} \Rightarrow \mathcal{N}(0, V_g) where NgN_g is group size and VgV_g is the appropriate sandwich covariance (Chu, 2017).

6. Regularity Assumptions

Establishing these properties relies on:

  • (A1) Stationarity and α\alpha-mixing: {(xit,uit)}\{(x_{it}, u_{it})\} is strictly stationary in tt and α\alpha-mixing across (i,t)(i, t) with mixing coefficients decaying sufficiently fast.
  • (A2) Group-heterogeneity: Each group contains a non-negligible share of units, and group-specific parameters are separated by at least δ>0\delta > 0 in norm.
  • (A3) Spatial-error process: (IW)(I-W) is invertible, and the spatial process is \ell-mixing in cross-section.
  • (A4) Moments: Uniform fourth moment bounds on xitx_{it} and uitu_{it}; parameters contained in a compact set (Chu, 2017).

7. Finite-Sample Behavior and Simulation Results

Monte Carlo experiments with N={100,200}N = \{100, 200\}, T={20,50}T = \{20, 50\}, and G={2,3}G = \{2, 3\} indicate:

  • Classification accuracy of unit grouping exceeds 95% even at moderate N,TN, T.
  • Bias and RMSE of (βg,ϕg,θg)(\beta_g, \phi_g, \theta_g) decrease at a (NT)1/2(NT)^{-1/2} rate, consistent with theory.
  • The CQL-DCA solution outperforms both naive pooled QMLE (ignoring groups) and pre-clustering (K-means plus two-step QMLE) on both parameter estimation bias and group assignment accuracy.

The empirical findings substantiate the theoretical properties, demonstrating both robustness to fixed effect misspecification and the computational viability of the overall estimation framework (Chu, 2017).


Chu's CQL methodology, by constructing a pairwise quasi-likelihood, embedding a mixed-integer group allocation, and applying DC algorithmic optimization, provides a rigorous, computationally tractable, and asymptotically efficient solution to high-dimensional, group-heterogeneous, spatially dependent dynamic panel models (Chu, 2017).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Composite Quasi-Likelihood (CQL).