Seemingly Unrelated Regression (SUR)
- Seemingly Unrelated Regression (SUR) is a system of linear models that allows for contemporaneously correlated errors across equations, enhancing estimation efficiency.
- It employs methods like generalized least squares (GLS) and feasible GLS, ensuring robust and efficient parameter estimation even with heterogeneous regressor sets.
- Recent extensions include Bayesian variable selection, high-dimensional precision estimation, and robust techniques that address measurement error and outlier contamination.
A seemingly unrelated regression (SUR) model is a multivariate system of linear regression equations in which each equation potentially has a different set of regressors, but the disturbance terms across equations for each observational unit are allowed to be contemporaneously correlated. This cross-equation correlation, while lacking economic or structural interdependence in the regressors, creates potential efficiency gains by joint estimation of the system rather than separate estimation of each equation. The SUR models—formally introduced by Zellner (1962)—form the backbone of multivariate regression analysis in econometrics, systems biology, finance, and related fields, and provide a general framework for enrichment with Bayesian selection, robust estimation, high-dimensional covariance learning, measurement error, and nonlinearity.
1. Mathematical Structure and Likelihood
The prototypical SUR system, with equations and observations, writes for and : or in matrix notation for each equation: with the response, the regressor matrix, the coefficient vector, and the error.
Stacking all equations: we have
with possessing a block covariance , where is the contemporaneous covariance matrix of the errors across equations.
Under Gaussianity , the log-likelihood is
The efficient generalized least squares (GLS) estimator is
with Feasible GLS (FGLS) obtained by plugging in a consistent estimator , usually based on OLS residuals (Saraceno et al., 2021, Peremans et al., 2018, Wali et al., 2019).
2. Identification, Covariance Structure, and Efficiency
Identification of requires full column rank of the joint design matrix, as well as non-singularity of the error covariance structure in the sense of the information matrix having rank equal to the parameter dimension, where is the full block covariance (possibly singular) and denotes Moore–Penrose inverse (Haupt, 2020). Restrictions may also arise explicitly from parameter constraints or implicitly from the covariance structure (e.g., zero-variance directions corresponding to adding-up or symmetry).
For panel or multi-equation models with time and group effects, the error often receives an additive decomposition: with group effects , period effects , and idiosyncratic terms. Variances of each component can be heterogeneous across strata (Platoni et al., 2016). Estimation proceeds by moment-based or ANOVA-style decomposition and system-wide GLS.
For generalizations such as singular covariance, explicit formulae for the constrained GLS estimator utilize the information from both the design and the covariance's low-rank representation; in fixed effects panel models, the within-estimator is a special case of singular SUR GLS (Haupt, 2020).
The semiparametric efficiency bound for is attained by feasible GLS, as shown via influence-function calculations, and coincides with the parametric GLS bound under mild conditions (Hristache et al., 2011).
3. Extensions: Bayesian Selection, High-dimensionality, Robustness
Bayesian SUR and Variable Selection: Modern SUR implementations, exemplified by the BayesSUR framework, introduce joint variable and covariance selection using priors such as spike-and-slab, "hotspot," and Markov random field (MRF) structures on support matrices and decomposable hyper-inverse-Wishart priors on (Zhao et al., 2021). Covariance factorization yields computationally scalable MCMC. Empirical evaluations demonstrate near-perfect recovery in high-dimensional genomics (eQTL/mQTL) and improved predictive density in applied drug-response tasks by incorporating structured priors on inclusion indicators. The modular approach allows full decoupling and recombination of variable- and covariance-selection architectures.
High-dimensional SUR: When the number of equations is comparable to or exceeds the number of observations , sample estimators of are ill-posed. The FGLasso estimator replaces with a sparse-precision estimator , exploiting graphical lasso regularization assumptions. When is sparse (banded, lattice), FGLasso achieves asymptotic efficiency equivalent to infeasible GLS; performance gains are demonstrated for (Tan et al., 2018).
Robust Estimation: Classical SUR estimators are highly sensitive to outliers, both at row- and cell-level. MM-estimators deliver high-breakdown, high-efficiency inference (50% breakdown achievable in the S-step; asymptotic coverage and power maintained in simulations). Fast robust bootstrap (FRB) procedures enable valid inference and hypothesis testing with substantial speedups over classical resampling (Peremans et al., 2018). Similarly, the two-step "surerob" estimator combines cellwise- and rowwise-robustification by an initial univariate MM step and multivariate 2SGS estimator for the residual covariance. Under independent cellwise contamination (ICM), surerob maintains nonzero breakdown and outperforms fastSUR and classical methods even as the number of equations grows; a plausible implication is that only methodologies explicitly targeting both contamination regimes should be applied in high-dimensional, multivariate settings (Saraceno et al., 2021).
The robustness–efficiency tradeoff and performance under distinct contamination schemes are summarized in the table below:
| Method | Breakdown (THCM) | Breakdown (ICM) | Computational Cost |
|---|---|---|---|
| Classical SUR | 0 | 0 | |
| MM-estimator | Up to 50% | 0 for large | High |
| FastSUR [Hubert+] | 50% | 0 for large | |
| surerob | 50% | , moderate | Moderate (dominated by MM and 2SGS) |
Non-Gaussian Errors and Mixture Models: SUR systems with non-normal error structure can be modeled via finite Gaussian mixtures. Identifiability is preserved under full-rank regressors. Likelihood, score, and Hessian are available in closed form, and ML estimation proceeds via EM. Empirical applications confirm improved fit and flexibility over classical SUR (Galimberti et al., 2014).
4. Model Selection, Measurement Error, and Variable Uncertainty
Model selection and predictor uncertainty: Modern approaches decouple statistical inference from variable selection using post-inference summarization and utility minimization. A penalized loss (including expected predictive accuracy and or penalties) is minimized, with summary coefficients computed via lasso or related convex optimization. This framework supports uncertainty in predictors (e.g., forecasting contexts), yielding sparser models compared to the fixed-X paradigm. Empirical recovery of meaningful factor structures in asset-pricing demonstrates the gain in deploying selection criteria tied to predictive information, not just marginal likelihood (Puelz et al., 2016).
Measurement error: Extensions to SUR with classical measurement error employ fully Bayesian algorithms—Gibbs sampling or mean-field variational Bayes—to infer both noise and regression parameters. Identification is secured by priors on measurement error variance. MFVB achieves comparable accuracy at a small fraction of the computational cost of MCMC. Ignoring measurement error biases both coefficients and covariance estimates, while the Bayesian approaches correct these biases and improve predictive fit (Bresson et al., 2020).
5. Nonlinearities, Structured Graphs, and Application Domains
Nonlinear multivariate regression can be embedded in a SUR context via spline-basis expansions and spike-and-slab priors on inclusion indicators, with the error precision endowed with a Gaussian graphical model prior (e.g., hyper-inverse-Wishart for decomposable graphs). Collapsing over regression coefficients and covariance enables efficient marginal MCMC for joint predictor–graph selection. Posterior model probabilities are consistent under increasing sample size, with the hyperparameters and design regularity controlling finite-sample behavior (Niu et al., 2020).
Applications of SUR models are broad, including local projection systems (where SUR likelihood arises naturally under VARMA data-generating processes) (Tanaka, 2020), multivariate forecasting, gene–drug association studies (Zhao et al., 2021), personalized fuel economy analysis using bivariate random-coefficient SUR (Wali et al., 2019), and stratified two-way panel models with SUR structure and heteroscedasticity (Platoni et al., 2016).
6. Summary Table of Modern SUR Generalizations
| Extension | Key Features | Reference |
|---|---|---|
| Bayesian variable/covariance selection | Hotspot/MRF priors on inclusion, HIW on , fast MCMC | (Zhao et al., 2021) |
| High-dimensional precision | FGLasso estimator, glasso on , large , sparse graph | (Tan et al., 2018) |
| Robust estimation | MM and S estimators; surerob for cell/row contamination | (Peremans et al., 2018, Saraceno et al., 2021) |
| Non-Gaussian error mixtures | Finite Gaussian mixtures, EM estimation, identifiability conditions | (Galimberti et al., 2014) |
| Measurement error | Bayesian MCMC/MFVB, prior on variance component | (Bresson et al., 2020) |
| Model selection under predictor uncertainty | Utility-based selection, post-inference lasso summarization | (Puelz et al., 2016) |
| Stratified heteroscedastic panels | Population/time stratification, component ANOVA estimation | (Platoni et al., 2016) |
7. Computational and Practical Considerations
Effective computation with SUR models depends on:
- Covariance factorization ( in block or graphical form) to avoid large matrix inversion, especially in MCMC (Zhao et al., 2021).
- Parallelization of likelihood calculation across CPUs for Bayesian methods.
- Algorithmic advances, such as Thompson-sampling-inspired "bandit" proposals for indicator matrices, and block-coordinate optimization in the graphical lasso (Tan et al., 2018).
- The trade-off between breakdown point and efficiency in robust approaches, and the choice of weight functions in MM or MM+S estimation (Peremans et al., 2018, Saraceno et al., 2021).
Advances in this domain have enabled robust, scalable, and interpretable multivariate analysis spanning genomics, economics, engineering, and environmental domains, with continuing innovation in covariance modeling, regularization, and integration of contextual structural priors.