Subvector AR Test Statistic
- Subvector AR test statistic is a generalized Anderson-Rubin method that profiles out nuisance parameters to focus on parameters of interest.
- The method minimizes a quadratic form over nuisance parameters, ensuring valid size control even under weak identification and heteroskedasticity.
- Enhanced with data-dependent critical values and regularization techniques, the approach yields reliable confidence intervals in both IV and time series models.
A subvector Anderson-Rubin (AR) test statistic is a set of procedures for valid hypothesis testing and confidence set construction on one or more components of a parameter vector in models with nuisance parameters, weak identification, and potential nonlinearity or heteroskedasticity. Subvector AR methodology generalizes the classical AR test by minimizing or profiling out nuisance parameters, yielding tests and intervals with guaranteed size control, often under minimal assumptions and for a wide array of identification regimes.
1. Core Definitions and Framework
The canonical subvector AR test addresses the hypotheses
where is a subvector of interest from the full parameter vector , with comprising nuisance parameters (Lounis, 2013). In time series AR() models,
one partitions accordingly and exploits the local asymptotic normality (LAN) expansion to form the test.
In linear instrumental variable (IV) models,
where are the parameters of interest and the nuisance parameters, the AR null tested is .
The subvector AR statistic operates by minimizing a criterion (typically a quadratic or generalized quadratic form in appropriate moments) over the nuisance parameter space, delivering a test statistic and confidence set that profile out the nuisance parameter and avoid reliance on possibly unidentified or weakly identified nuisance components (Londschien et al., 2024, Inoue et al., 1 Jul 2025).
2. Mathematical Formulation of the Subvector AR Statistic
Time Series Models
For AR() processes, the subvector test statistic is constructed as
where is the efficient score for , obtained by removing the effect of the nuisance score (via regression and the block-Fisher information matrix ) (Lounis, 2013):
with
Under , .
Linear IV Model
For a model
with as instruments, the subvector AR statistic (profiled in ) is (Londschien et al., 2024):
where , , and . The minimizer corresponds to the LIML-type fit; plugging it in yields a closed-form test statistic.
In the most general GMM/M-estimation setting, the subvector AR statistic can be written as
with
where is the sample mean of the moment conditions and a HAC or robust covariance estimator (Inoue et al., 1 Jul 2025).
3. Distributional Theory and Critical Values
Across models and implementations, the subvector AR test is designed such that, under regularity conditions and (regardless of identification strength),
with the dimension of the subvector of interest, the "effective" degrees of freedom after profiling out nuisance parameters, and the dimension of .
A critical innovation is the use of data-dependent critical values, particularly in the IV setting with potentially weak instruments. Guggenberger, Kleibergen and Mavroeidis (GKM) propose conditional critical values—tabulated as —functioning of a conditioning variable , typically the largest or second-smallest eigenvalue of a matrix formed from the data (Hoekstra et al., 25 Jan 2026). When , conditioning on the second-smallest eigenvalue yields strictly higher power while maintaining exact or conservative size:
In the general GMM/CUE setting, no data-dependent critical values or bootstrapping are required: one uses the nominal cut-off for uniformly valid subvector inference (Inoue et al., 1 Jul 2025).
4. Robustness to Identification, Heteroskedasticity, and High Dimensionality
The subvector AR approach is explicitly designed to be robust in settings of weak, partial, or non-identification of nuisance parameters:
- In linear IV models, the AR test achieves weak-instrument-robust size under minimal moment and rank conditions (Londschien et al., 2024).
- In GMM and nonlinear moment models, uniform validity is achieved by perturbing the first-order CUE conditions, employing truncated SVD for regularization, and ensuring all identification regimes are covered (Inoue et al., 1 Jul 2025).
- Heteroskedasticity-robust variants use kernel or block-diagonal (AKP) covariance estimators, maintaining validity and typically improving power when the structure is appropriately captured (Guggenberger et al., 2021, Hoekstra et al., 25 Jan 2026).
Modern enhancements deploy ridge regularization and jackknife bias correction (RJAR), extending subvector AR tests to cases where the number of instruments exceeds sample size or covariance matrices are ill-conditioned, while maintaining weak-IV and heteroskedasticity robustness (Dovì et al., 2022).
5. Power Properties and Optimality
Subvector AR tests exhibit several notable optimality and power properties:
- For local alternatives, the efficient-score-based AR test in LAN models is locally most powerful invariant and attains the Le Cam optimality bound (Lounis, 2013).
- Conditioning critical values on the second-smallest eigenvalue (instead of the largest) results in uniformly higher power for , since the rejection region more accurately adapts to the weakest direction of identification (Hoekstra et al., 25 Jan 2026).
- In the heteroskedastic AKP setting, model selection ensures that validity is always maintained, and power gains accrue when the covariance exhibits approximate Kronecker structure (Guggenberger et al., 2021).
- Simulations confirm that ridge-regularized and jackknifed AR tests outperform their nonregularized analogs in high dimensional, many-instrument regimes (Dovì et al., 2022).
6. Inference: Confidence Regions and Implementation
Confidence sets are formed by inversion of the AR test, either as the set of subvector parameter values for which the test statistic does not exceed the critical value, or, for scalar subvector problems, by searching for roots of the AR test equation. In the linear IV framework, the confidence set can be written explicitly as a (possibly unbounded) ellipsoid centered at a k-class estimator, with precise closed-form dependence on the data and critical value (Londschien et al., 2024).
Implementation for both linear and nonlinear models involves minimizing the AR criterion over nuisance parameters, computation of HAC or block-diagonal covariance matrices as applicable, and lookup or calculation of critical values. No bootstrapping is required for nominal size control, provided the model-specific distributional results hold (Inoue et al., 1 Jul 2025, Hoekstra et al., 25 Jan 2026). For high dimensional or weakly identified settings, special care in optimization and estimation (jackknife, ridge, Kronecker estimates) is warranted (Dovì et al., 2022, Guggenberger et al., 2021).
7. Extensions and Practical Recommendations
- The methodology extends to general time series (including nonstationary, nonlinear, and heteroskedastic processes), cross-sectional models, and cases with more instruments than observations.
- When the covariance structure closely follows a Kronecker product, AKP-based tests are recommended for power; when this structure fails, fully robust AR procedures are preferred for exactness (Guggenberger et al., 2021, Hoekstra et al., 25 Jan 2026).
- For multiple nuisance parameters, conditioning on the second-smallest eigenvalue is recommended for test power and size (Hoekstra et al., 25 Jan 2026).
- Ridge regularization is essential in "many instrument" settings to ensure well-posedness and valid inference (Dovì et al., 2022).
Summary Table: Subvector AR Statistic Variants
| Model Context | Test Statistic / Conditioning | Robustness |
|---|---|---|
| AR() time series | Efficient score quadratic form | LAN, optimal, reparam.-inv |
| Linear IV, homoskedastic | , cond. on | Weak IV, strong IV |
| Linear IV, heteroskedastic | AKP-AR, switching to AR/AR | Arbitrary heteroskedasticity |
| Nonlinear GMM/CUE | Prof. over | Weak/partial/non-ID, HAC |
| High-dimensional IV | Ridge-jackknife AR | , weak IV, het. |
All methods aim to preserve correct asymptotic size, provide uniform validity over identification regimes, and exploit conditioning or regularization to maximize power, particularly in models with many or weak instruments and/or heteroskedasticity (Lounis, 2013, Londschien et al., 2024, Hoekstra et al., 25 Jan 2026, Inoue et al., 1 Jul 2025, Dovì et al., 2022, Guggenberger et al., 2021).