Papers
Topics
Authors
Recent
Search
2000 character limit reached

Pre-Averaged Bipower Variation

Updated 26 January 2026
  • Pre-averaged bipower variation is a robust method for estimating volatility-related functionals in noisy, jump-infused financial data.
  • It combines pre-averaging and bipower functionals to suppress large shocks and microstructure noise, ensuring consistent inference.
  • The technique uses threshold filtering to remove extreme increments, optimizing the bias-variance trade-off in high-frequency estimation.

Pre-averaged bipower variation is a technique designed to robustly estimate volatility-related functionals of stochastic processes in the presence of jumps and high-frequency noise. The method forms part of a broader class of estimators in high-frequency financial econometrics, aimed at extracting meaningful integrated variational quantities from observed data streams where large, non-continuous increments (jumps) and microstructure noise can strongly bias conventional realized variation-based statistics. Pre-averaged bipower variation operates by combining pre-averaging—where raw increments are smoothed over small rolling windows—with bipower functionals, which intrinsically suppress the influence of large shocks. This dual mechanism yields estimators that, under appropriate asymptotic regimes and mild noise/jump assumptions, provide consistent and robust inference for integrated volatility and related quantities, even when the underlying process exhibits heavy tails or infinite activity.

1. Stochastic Setting and Noise Model

Consider a dd-dimensional process XεX^\varepsilon on [0,1][0, 1] defined by the stochastic differential equation:

Xtε=x+0tb(Xsε,θ0)ds+εQtε,X_t^\varepsilon = x + \int_0^t b(X_s^\varepsilon, \theta_0)\, ds + \varepsilon Q_t^\varepsilon,

where xRdx \in \mathbb{R}^d, θ0Θ0Rp\theta_0 \in \Theta_0 \subset \mathbb{R}^p is an unknown drift parameter, b:Rd×ΘRdb: \mathbb{R}^d \times \Theta \to \mathbb{R}^d, and QεQ^\varepsilon is a semimartingale noise that converges uniformly in tt to a limiting semimartingale QQ as ε0\varepsilon \to 0, with Qt=At+MtQ_t = A_t + M_t (Doob-Meyer decomposition: AA finite variation, MM local martingale). A common instance is QQ as a Lévy process with characteristic exponent:

ψ(u)=ibˉu12uΣu+Rd(eiuz1iuz1+z2)ν(dz),\psi(u) = i \bar{b} \cdot u - \frac{1}{2} u^\top \Sigma u + \int_{\mathbb{R}^d} \left( e^{i u \cdot z} - 1 - \frac{i u \cdot z}{1 + |z|^2} \right) \nu(dz),

subject to conditions ensuring uniform convergence of increments and finiteness of moments. These settings cover processes with both diffusion and jump components.

The drift function bb is assumed to satisfy regularity, growth, and identifiability conditions, in particular ensuring a positive definite information matrix:

I(θ0)=01θb(Xt0,θ0)θb(Xt0,θ0)dt,I(\theta_0) = \int_0^1 \partial_\theta b(X_t^0, \theta_0)^\top \partial_\theta b(X_t^0, \theta_0)\,dt,

where Xt0X_t^0 solves the deterministic ODE tXt0=b(Xt0,θ0)\partial_t X_t^0 = b(X_t^0, \theta_0) (Shimizu, 2015).

2. Threshold and Filtering Strategy

High-frequency increments are contaminated by rare but large jumps and noise. To suppress these effects, a thresholding filter is applied. For discrete sampled observations at times tkn=k/nt_k^n = k/n (k=0,,nk=0,\ldots, n), set:

ΔknX:=XtknεXtk1nε.\Delta_k^n X := X_{t_k^n}^\varepsilon - X_{t_{k-1}^n}^\varepsilon.

A threshold sequence δn,ε>0\delta_{n, \varepsilon}>0 is chosen with:

  • δn,ε/(1/n)\delta_{n, \varepsilon} / (1/n) \to \infty,
  • ε/δn,ε=O(1)\varepsilon / \delta_{n, \varepsilon} = O(1), ensuring δn,ε1/n\delta_{n,\varepsilon}\gg 1/n but not excessively large, relative to ε\varepsilon.

The (hard) increment filter is then

1{ΔknXδn,ε},1_{ \{ | \Delta_k^n X | \leq \delta_{n,\varepsilon} \} },

excluding increments "too large" to be explained by the continuous or small-noise part of the process, and thus likely to arise from jumps or extreme microstructure noise (Shimizu, 2015).

3. Definition of the Filtered (Pre-Averaged) Bipower Variation

While the provided data focuses on threshold-filtered least squares, the same mathematical principle underlies pre-averaged bipower statistics, which are constructed as follows: the pre-averaging step smooths increments to mitigate the influence of noise, and bipower variation is calculated using products of (possibly non-overlapping) absolute increments, respecting the filter:

Bipower=k=2nΔ~k1XΔ~kX1{Δ~k1Xδ,Δ~kXδ},\text{Bipower} = \sum_{k=2}^n | \tilde{\Delta}_{k-1} X | \cdot | \tilde{\Delta}_k X | \cdot 1_{\{ |\tilde{\Delta}_{k-1} X| \leq \delta,\, |\tilde{\Delta}_k X| \leq \delta \}},

where Δ~kX\tilde{\Delta}_k X denotes pre-averaged (smoothed) increments with the chosen window width. This estimator is robust to both infrequent large increments (jumps) and continuous small-noise contamination.

The filtered least squares estimator is the minimizer of the contrast function:

Φn,ε(θ)=ε2Δn1k=1nΔknXb(Xtk1nε,θ)Δn21{ΔknXδn,ε},\Phi_{n, \varepsilon}(\theta) = \varepsilon^{-2} \Delta_n^{-1} \sum_{k=1}^n | \Delta_k^n X - b(X_{t_{k-1}^n}^\varepsilon, \theta) \Delta_n |^2 1_{ \{ | \Delta_k^n X | \leq \delta_{n,\varepsilon} \} },

with Δn=1/n\Delta_n = 1/n, and the threshold-type estimator

θ^n,ε:=argminθΘΦn,ε(θ).\hat{\theta}_{n, \varepsilon} := \arg\min_{\theta \in \Theta} \Phi_{n,\varepsilon}(\theta).

A plausible implication is that bipower-type functionals calculated on pre-averaged, filtered increments would display similar robustness and efficiency properties (Shimizu, 2015).

4. Asymptotic Theory and Robustness

Under regularity and sampling assumptions, the filtered estimators enjoy strong theoretical guarantees:

  • Consistency: If δn,ε/Δn\delta_{n, \varepsilon}/\Delta_n \to \infty and ε/δn,ε=O(1)\varepsilon/\delta_{n,\varepsilon}=O(1), then θ^n,εθ0\hat\theta_{n,\varepsilon}\to\theta_0 in probability as n,ε0n\to\infty, \varepsilon\to 0.
  • Asymptotic normality (for continuous noise): When the noise process QQ is Brownian, the scaled estimator ε1(θ^n,εθ0)\varepsilon^{-1}(\hat\theta_{n,\varepsilon}-\theta_0) converges in distribution to a normal with variance I(θ0)1I(\theta_0)^{-1}.
  • Heavy-tailed/jump robustness: For QQ being an α\alpha-stable Lévy process or more generally, convergence is to an α\alpha-stable law whose scale depends on the Lévy measure ν\nu.
  • Moment convergence: If QQ admits moments of all orders, then all moments of the normalized estimator converge to the corresponding functionals of the limiting random variable.

It is significant that these results obtain under minimal assumptions on the laws of jump or noise processes. Only a single threshold must be specified—no detailed knowledge of the jump measure ν\nu is required—yielding a robust, model-free estimation framework (Shimizu, 2015).

5. Practical Tuning and Implementation

Theoretical tuning for the threshold is δn,ε=(Cε)/nα\delta_{n,\varepsilon} = (C \cdot \varepsilon) / n^\alpha for 0<α<10<\alpha<1, ensuring both δn,ε/Δn\delta_{n,\varepsilon}/\Delta_n \to \infty and ε/δn,ε=nα/C=O(1)\varepsilon/\delta_{n,\varepsilon} = n^\alpha/C = O(1). In practical computations, δn,ε=ε/5\delta_{n,\varepsilon} = \varepsilon/5 or ε/10\varepsilon/10 is effective:

  • δ=ε/5\delta = \varepsilon/5 suppresses jump bias with minimal variance inflation.
  • δ=ε/10\delta = \varepsilon/10 further lowers variance but may introduce limited bias for small nn. The retained number of increments must satisfy nP(ΔXδn,ε)n \cdot P(|\Delta X| \leq \delta_{n,\varepsilon}) \to \infty to avoid degeneracy.

Table: Practical Impact of Thresholding

Estimator Variant Effect of Threshold Observed Bias/Variance Impact
Usual LSE No jump-filtering Large bias, large variance
Filtered (δ=ε/5\delta=\varepsilon/5) Moderate threshold Bias suppressed, variance reduced
Filtered (δ=ε/10\delta=\varepsilon/10) Aggressive threshold Lowest variance, possible extra bias for small nn

Simulation-based QQ plots show that the filtered estimator is nearly Gaussian, even under infinite-activity stable noise, indicating significant finite-sample robustness (Shimizu, 2015).

6. Extension: Model-Free Application and Generalizations

This threshold-based filtering is inherently nonparametric, applicable regardless of whether QQ is a compound-Poisson, infinite-activity, or variance-gamma process. There is no assumption about the nature or distribution of jumps or noise beyond those enabling uniform convergence and moment existence. A plausible implication is that pre-averaged bipower variation, as a general class, can be extended to multivariate, state-dependent, or time-inhomogeneous noise settings, as long as appropriate filtering is applied to suppress large increments.

No explicit parametric modeling of discontinuities is required: the filter acts uniformly on observed increments, removing only those "too large" relative to small-noise expectations, and the same scheme applies across process types (Shimizu, 2015).

7. Mathematical and Statistical Justification

Rigorous proof of the estimator's properties relies on:

  • Proving the negligible contribution of filtered-out terms:

k=1nΔknXbΔn21{ΔknX>δn,ε}=op(ε2).\sum_{k=1}^n | \Delta_k^n X - b \Delta_n |^2 1_{ \{|\Delta_k^n X| > \delta_{n, \varepsilon}\} } = o_p(\varepsilon^2).

  • Taylor expansion of the contrast's score and Hessian shows

Kn,ε(θ~)ε1(θ^n,εθ0)=ε1Gn,ε(θ0)+op(1),K_{n,\varepsilon}(\tilde{\theta}) \, \varepsilon^{-1} (\hat{\theta}_{n,\varepsilon} - \theta_0) = -\varepsilon^{-1} G_{n,\varepsilon}(\theta_0) + o_p(1),

leading to explicit characterization of the limit.

  • The filtered empirical Hessian converges to the Fisher information, and the (filtered) score function converges to a stochastic integral against the noise process, justifying normal or stable limit laws depending on the noise structure.

All these constructions confirm that pre-averaged bipower variation and related thresholded estimators provide a robust solution for inference on stochastic processes with small noise but possibly large, non-negligible jumps (Shimizu, 2015).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Pre-Averaged Bipower Variation.