Order Parameter Bootstrapping in Model Selection
- Order parameter bootstrapping is a statistical framework that combines Lepski's method, multiple testing, and wild bootstrap calibration to control type I error.
- It rigorously selects the optimal bias-variance trade-off in linear estimators while achieving oracle-like risk bounds in finite samples.
- The method leverages pairwise test statistics and tailored thresholds to adaptively manage heteroscedastic noise without relying on strict parametric assumptions.
The order parameter bootstrapping argument is a statistical framework for model selection within a family of linear estimators, with a focus on controlling type I error in a completely data-driven manner that is robust to unknown and potentially heteroscedastic noise. The method, developed primarily in the context of ordered model selection for smoothing estimators, advances the "smallest accepted" approach, integrating Lepski's method with contemporary multiple testing theory and leveraging wild (multiplier) bootstrap procedures for threshold calibration. The result is a rigorous, finite-sample-valid, heteroscedasticity-robust procedure for selecting the optimal point along a sequence of bias-variance trade-offs, equipped with provable oracle‐type risk bounds and explicit error control (Spokoiny et al., 2015).
1. Framework: Ordered Linear Estimators and Model Selection
The foundational setup involves observed data , where is an unknown signal and is a zero-mean noise vector with covariance , which may be heteroscedastic. Candidate estimators are constructed as for , with fixed smoothing matrices . These estimators are ordered in terms of variance: . Each model in the sequence offers a specific bias-variance trade-off, with lower- models generally incurring higher bias and lower variance, and higher- models the opposite. An acceptance rule formalizes statistical indistinguishability: model is accepted if it cannot be reliably distinguished from all larger models , based on the observed data.
2. Pairwise Test Statistics and Acceptance Thresholds
For each pair with , the difference is considered, with covariance . Under the null hypothesis that model is true ( in the column space of ), this difference is zero-mean with covariance . The primary diagnostic is the quadratic form
where is a suitable generalized inverse, and under the null, is approximately -distributed with degrees of freedom. A model is accepted if for all , where are critical values selected to control the family-wise error rate of false rejections via the propagation condition.
3. Wild Bootstrap Calibration for Threshold Selection
A central challenge is that is both unknown and potentially heteroscedastic. The method addresses this through a multiplier (wild) bootstrap to approximate the joint distribution of all test statistics under each model :
- A pilot estimator (often ) defines initial residuals .
- For bootstrap iterations, i.i.d. mean-zero multipliers with are generated (e.g., Rademacher or standard normal).
- Bootstrap pseudo-observations are constructed as .
- The test statistics are recomputed for all on the bootstrapped data.
- The empirical distribution of is used to set , where the levels are chosen so that , with the desired overall error level.
This procedure ensures that for any true model , the probability of any false rejection among for is bounded by .
4. Selection Rule and Type I Error Control
The final selection rule is
Equivalently, the process begins at the lowest index and selects the smallest model not rejected by any larger model. By the construction of and the bootstrap-based thresholds, it follows that, for any in the column space of some ,
where and are constants dependent on the Monte-Carlo and non-Gaussian tail behavior of the test statistics, respectively. This upper bound demonstrates finite-sample family-wise error control for under-smoothing.
5. Finite-Sample Validity and Oracle Inequalities
Under suitable conditions (fixed design, sub-Gaussian noise, moderate dimensionality; see assumptions A1–A3 in the original work), the method guarantees not only type I error control but also risk performance nearly matching the oracle model. Specifically, with probability at least ,
where is a negligible remainder. The result holds equally for full parameter estimation or for linear functionals. This demonstrates that the method is data-adaptive and achieves a nearly optimal bias-variance balance, subject only to small finite-sample corrections (Spokoiny et al., 2015).
6. Methodological Context and Significance
The order parameter bootstrapping argument builds directly on Lepski's method and the theory of multiple testing, operationalized in a model selection context through the "smallest accepted" approach. Its primary contributions are the rigorous bootstrapped calibration of critical values in the presence of heteroscedastic noise, the explicit finite-sample family-wise error control, and the explicit characterization of the selected index set. The framework is fully data-driven, requiring neither prior variance structure nor parametric assumptions on the noise. These features position it as a robust alternative to classical penalized criteria, particularly in scenarios with non-constant variance and complex dependence structures in the residuals (Spokoiny et al., 2015).