Pre-Averaged Bipower Variation
- Pre-averaged bipower variation is a robust method for estimating volatility-related functionals in noisy, jump-infused financial data.
- It combines pre-averaging and bipower functionals to suppress large shocks and microstructure noise, ensuring consistent inference.
- The technique uses threshold filtering to remove extreme increments, optimizing the bias-variance trade-off in high-frequency estimation.
Pre-averaged bipower variation is a technique designed to robustly estimate volatility-related functionals of stochastic processes in the presence of jumps and high-frequency noise. The method forms part of a broader class of estimators in high-frequency financial econometrics, aimed at extracting meaningful integrated variational quantities from observed data streams where large, non-continuous increments (jumps) and microstructure noise can strongly bias conventional realized variation-based statistics. Pre-averaged bipower variation operates by combining pre-averaging—where raw increments are smoothed over small rolling windows—with bipower functionals, which intrinsically suppress the influence of large shocks. This dual mechanism yields estimators that, under appropriate asymptotic regimes and mild noise/jump assumptions, provide consistent and robust inference for integrated volatility and related quantities, even when the underlying process exhibits heavy tails or infinite activity.
1. Stochastic Setting and Noise Model
Consider a -dimensional process on defined by the stochastic differential equation:
where , is an unknown drift parameter, , and is a semimartingale noise that converges uniformly in to a limiting semimartingale as , with (Doob-Meyer decomposition: finite variation, local martingale). A common instance is as a Lévy process with characteristic exponent:
subject to conditions ensuring uniform convergence of increments and finiteness of moments. These settings cover processes with both diffusion and jump components.
The drift function is assumed to satisfy regularity, growth, and identifiability conditions, in particular ensuring a positive definite information matrix:
where solves the deterministic ODE (Shimizu, 2015).
2. Threshold and Filtering Strategy
High-frequency increments are contaminated by rare but large jumps and noise. To suppress these effects, a thresholding filter is applied. For discrete sampled observations at times (), set:
A threshold sequence is chosen with:
- ,
- , ensuring but not excessively large, relative to .
The (hard) increment filter is then
excluding increments "too large" to be explained by the continuous or small-noise part of the process, and thus likely to arise from jumps or extreme microstructure noise (Shimizu, 2015).
3. Definition of the Filtered (Pre-Averaged) Bipower Variation
While the provided data focuses on threshold-filtered least squares, the same mathematical principle underlies pre-averaged bipower statistics, which are constructed as follows: the pre-averaging step smooths increments to mitigate the influence of noise, and bipower variation is calculated using products of (possibly non-overlapping) absolute increments, respecting the filter:
where denotes pre-averaged (smoothed) increments with the chosen window width. This estimator is robust to both infrequent large increments (jumps) and continuous small-noise contamination.
The filtered least squares estimator is the minimizer of the contrast function:
with , and the threshold-type estimator
A plausible implication is that bipower-type functionals calculated on pre-averaged, filtered increments would display similar robustness and efficiency properties (Shimizu, 2015).
4. Asymptotic Theory and Robustness
Under regularity and sampling assumptions, the filtered estimators enjoy strong theoretical guarantees:
- Consistency: If and , then in probability as .
- Asymptotic normality (for continuous noise): When the noise process is Brownian, the scaled estimator converges in distribution to a normal with variance .
- Heavy-tailed/jump robustness: For being an -stable Lévy process or more generally, convergence is to an -stable law whose scale depends on the Lévy measure .
- Moment convergence: If admits moments of all orders, then all moments of the normalized estimator converge to the corresponding functionals of the limiting random variable.
It is significant that these results obtain under minimal assumptions on the laws of jump or noise processes. Only a single threshold must be specified—no detailed knowledge of the jump measure is required—yielding a robust, model-free estimation framework (Shimizu, 2015).
5. Practical Tuning and Implementation
Theoretical tuning for the threshold is for , ensuring both and . In practical computations, or is effective:
- suppresses jump bias with minimal variance inflation.
- further lowers variance but may introduce limited bias for small . The retained number of increments must satisfy to avoid degeneracy.
Table: Practical Impact of Thresholding
| Estimator Variant | Effect of Threshold | Observed Bias/Variance Impact |
|---|---|---|
| Usual LSE | No jump-filtering | Large bias, large variance |
| Filtered () | Moderate threshold | Bias suppressed, variance reduced |
| Filtered () | Aggressive threshold | Lowest variance, possible extra bias for small |
Simulation-based QQ plots show that the filtered estimator is nearly Gaussian, even under infinite-activity stable noise, indicating significant finite-sample robustness (Shimizu, 2015).
6. Extension: Model-Free Application and Generalizations
This threshold-based filtering is inherently nonparametric, applicable regardless of whether is a compound-Poisson, infinite-activity, or variance-gamma process. There is no assumption about the nature or distribution of jumps or noise beyond those enabling uniform convergence and moment existence. A plausible implication is that pre-averaged bipower variation, as a general class, can be extended to multivariate, state-dependent, or time-inhomogeneous noise settings, as long as appropriate filtering is applied to suppress large increments.
No explicit parametric modeling of discontinuities is required: the filter acts uniformly on observed increments, removing only those "too large" relative to small-noise expectations, and the same scheme applies across process types (Shimizu, 2015).
7. Mathematical and Statistical Justification
Rigorous proof of the estimator's properties relies on:
- Proving the negligible contribution of filtered-out terms:
- Taylor expansion of the contrast's score and Hessian shows
leading to explicit characterization of the limit.
- The filtered empirical Hessian converges to the Fisher information, and the (filtered) score function converges to a stochastic integral against the noise process, justifying normal or stable limit laws depending on the noise structure.
All these constructions confirm that pre-averaged bipower variation and related thresholded estimators provide a robust solution for inference on stochastic processes with small noise but possibly large, non-negligible jumps (Shimizu, 2015).