Papers
Topics
Authors
Recent
Search
2000 character limit reached

Quantile Least Squares (Q-LS)

Updated 28 January 2026
  • Quantile Least Squares (Q-LS) is a family of estimators that match sample quantiles to model-implied quantiles using least squares, ensuring robust parameter estimation.
  • It applies to various models including location-scale, regression, and instrumental variables, achieving high asymptotic efficiency and accurate tail fitting.
  • Q-LS methods emphasize computational scalability and robustness with bounded influence functions, making them suitable for handling outliers in large datasets.

Quantile Least Squares (Q-LS) is a broad family of parametric and semiparametric statistical estimators where the fundamental principle is to match (possibly multiple) sample quantiles or quantile relationships by least squares or related norm-minimizing criteria. Methodologies under the Q-LS designation span robust parameter estimation in location-scale families, distributional modeling, instrumental variable analysis, and robust regression (including order-statistic-based Chebyshev fits), all utilizing quantile information as the modeling basis (Adjieteh et al., 2024, Peng et al., 2023, Bertsimas et al., 2013, Puerto et al., 2024, Cherodian et al., 23 Jan 2026).

1. Core Definitions and Mathematical Structure

Let X1,,XnX_1,\dots,X_n be i.i.d. samples, or, in the regression case, (yi,xi)(y_i, x_i), i=1,,ni=1,\dots,n. The Q-LS approach constructs an estimator by (i) selecting a set of quantile levels {p1,,pk}(0,1)\{p_1,\dots,p_k\}\subset (0,1), (ii) computing empirical quantiles Q^j\widehat Q_j at each level, and (iii) fitting model-implied quantiles—possibly parameterized, or as a linear combination of basis quantile functions—by least-squares or related loss.

Two canonical instances:

  • Location-scale family Q-LS estimation: Given a parametric family F(x;μ,σ)F(x;\mu,\sigma) with F1(p)=μ+σq0(p)F^{-1}(p) = \mu + \sigma q_0(p), define the "ordinary" Q-LS estimator (oQLS) as

(μ^,σ^)=argminμ,σ>0j=1k(Q^j[μ+σq0(pj)])2,(\widehat\mu, \widehat\sigma)' = \arg\min_{\mu, \sigma > 0} \sum_{j=1}^k (\widehat Q_j - [\mu+\sigma q_0(p_j)])^2,

and the "generalized" version (gQLS) as weighted LS with weights given by the quantile asymptotic covariance matrix (Adjieteh et al., 2024).

  • Quantile regression / Least Quantile of Squares (LQS): For (yi,xi)(y_i, x_i), the LQS estimator finds

β^LQS(τ)=argminβr(q)(β),\hat\beta^{\rm LQS}(\tau) = \arg\min_\beta |r_{(q)}(\beta)|,

where ri(β)=yixiTβr_i(\beta) = y_i-x_i^T\beta, q=τnq=\lceil \tau n \rceil, and r(q)|r_{(q)}| is the qqth smallest absolute residual (Bertsimas et al., 2013, Puerto et al., 2024). This connects to the Chebyshev fit on subsets of size qq.

Other Q-LS formulations model the quantile function as a linear combination of basis quantile functions G(p;θ)=i=0IθiQi(p)G(p;\theta) = \sum_{i=0}^I \theta_i Q_i(p), and estimate θ\theta by constrained least squares on the fitted quantiles (Peng et al., 2023). For instrumental variables, Q-LS projects XX onto a dictionary of conditional quantiles QXZ(τZ)Q_{X|Z}(\tau|Z), aggregating them with optimal mean-square weights for IV estimation (Cherodian et al., 23 Jan 2026).

2. Asymptotic Theory and Efficiency

Q-LS estimators rely on the joint asymptotic normality of order statistics and sample quantiles:

n(Q^jF1(pj))j=1kdNk(0,Σ),\sqrt{n} (\widehat Q_j-F^{-1}(p_j))_{j=1}^k \xrightarrow{d} N_k(0,\,\Sigma),

with Σij=pi(1pj)/[f(F1(pi))f(F1(pj))]\Sigma_{ij}=p_i(1-p_j)/[f(F^{-1}(p_i))f(F^{-1}(p_j))] for the univariate i.i.d. case (Adjieteh et al., 2024, Peng et al., 2023).

  • oQLS: Standard LS asymptotics apply, but with sub-optimal variance when quantile covariances are ignored.
  • gQLS: Achieves semiparametric efficiency by using the covariance structure for optimal weighting; in location-scale families, achieves 90–99% asymptotic efficiency relative to MLE for practical values k15k\approx 15–$25$ (Adjieteh et al., 2024).
  • Mixture Q-LS: Large-NN limit minimizes LqL_q-Wasserstein distance between true and fitted quantile functions, and, under regularity, Gaussian behavior for fixed quantile grids (Peng et al., 2023).

Robust LQS regression achieves a maximal breakdown point of (nk+1)/n(n-k+1)/n, e.g., 50% for least median of squares (Bertsimas et al., 2013, Puerto et al., 2024). The influence function for Q-LS estimators is bounded by the interior quantile trimming, yielding robustness controlled by trimming parameters.

3. Computational Algorithms and Scalability

Q-LS estimators are computationally tractable for moderate to very large sample sizes, with algorithmic complexity depending on specific formulation:

  • Location-scale Q-LS: For kk quantiles, total computational cost is O(n)O(n) for quantile extraction plus small matrix algebra; feasible up to n109n \sim 10^9 with negligible overhead (Adjieteh et al., 2024).
  • Mixture Q-LS: Reduces to quadratic programming with convex or linear constraints, solved efficiently even with up to thousands of basis quantiles, and readily regularizable (e.g., 1\ell_1, 2\ell_2 penalties) (Peng et al., 2023).
  • LQS regression: NP-hard in general (Puerto et al., 2024), but advances enable provable global solutions for n500n\lesssim 500 via modern MILP, bilevel, or sorting-based MIP reformulations, and subpercent optimality or high-quality solutions via continuous relaxation and warm-starts for nn into the low thousands (Bertsimas et al., 2013, Puerto et al., 2024). Aggregation heuristics cluster data for scalable approximate Q-LS in large-nn settings.
  • Efficient quantile regression: Recent combinatorial and randomized divide-and-conquer techniques achieve strongly polynomial algorithms in O(nlog2n)O(n\log^2 n) for 2D and O(nd1log2n)O(n^{d-1}\log^2 n) in higher dimensions, offering order-of-magnitude speedups over traditional LP or IP solvers for large nn (Shetiya et al., 2023).

4. Robustness and Influence Functions

Q-LS methodologies provide intrinsic robustification due to their quantile-centric nature:

  • For location-scale Q-LS, the influence function for both μ^\widehat\mu and σ^\widehat\sigma is bounded by construction via interior quantile trimming (excluding extremes [a,b](0,1)[a,b]\subset(0,1)), yielding a breakdown point min{a,1b}\min\{a,1-b\} (Adjieteh et al., 2024). The resulting influence functions resemble classic robust estimators: Tukey's biweight for location, and MM-scales for scale, but without explicit specification.
  • In regression, LQS estimators (e.g., least median, least trimmed squares) reach maximal breakdown for τ=0.5\tau=0.5 (median), demoting the effect of up to 50% contamination before breakdown (Bertsimas et al., 2013, Puerto et al., 2024). This is independent of dimensionality pp.
  • In the mixture-Q-LS context, monotonicity and tail constraints, as well as regularization penalties, can bolster finite-sample robustness (Peng et al., 2023).

5. Model Validation and Applications

Q-LS supports model checking and validation through quantile-based residuals and bespoke goodness-of-fit tests:

  • χ²-type in-sample validation (location-scale gQLS): Test statistic

W=nσ^2(YXβ^)Σ1(YXβ^)χk22,W = \frac{n}{\widehat\sigma^2}(\mathbf Y - \mathbf X\widehat\beta)'\Sigma_*^{-1}(\mathbf Y-\mathbf X\widehat\beta) \approx \chi^2_{k-2},

controlling level and showing high power for alternatives (Adjieteh et al., 2024).

  • Out-of-sample validation: Resampling on test quantile grids, using a parametric bootstrap to compute pp-values as the null is not analytic (Adjieteh et al., 2024).
  • Empirical illustration: Google daily stock returns (n=986) strongly prefer the logistic over normal, Laplace, Gumbel, or Cauchy under gQLS-based tests (Adjieteh et al., 2024).
  • Instrumental Variables: Q-LS improves weak-instrument situations by aggregating conditional quantiles for IV estimation, enabling consistent inference even when the conditional mean is flat but distributional relevance persists (Cherodian et al., 23 Jan 2026).
  • Distributional modeling: Q-LS mixture frameworks outperform MLE and are markedly superior at tail fitting for financial datasets, e.g., S&P500 drawdowns (Peng et al., 2023).

6. Implementation Guidelines and Parameter Choices

Practical implementation of Q-LS hinges on quantile grid selection, trimming, weighting, and computational solver choice:

  • For location-scale Q-LS, k=15k=15–$25$ equally spaced, interior quantiles (a,ba,b away from 0 and 1) are sufficient for maximal efficiency; a=0.05a=0.05, b=0.95b=0.95 is recommended for moderate robustness (Adjieteh et al., 2024).
  • Generalized Q-LS (gQLS) weighting is always preferable due to improved efficiency at negligible extra cost (Adjieteh et al., 2024, Peng et al., 2023).
  • In mixture Q-LS, monotonicity (quantile gridwise) and nonnegativity constraints ensure model validity, with regularization aiding parsimony and performance in high-dimensional basis spaces (Peng et al., 2023).
  • For regression LQS, the indicator-based bilevel reformulation is most efficient up to n100n\simeq 100–$200$ (Puerto et al., 2024). Aggregation heuristics cluster large nn to tractable size, with error bounded by cluster diameter.
  • For instrumental variables Q-LS, quantile grid size KK requires K2/n0K^2/n \to 0 for theoretical results; ridge or LASSO regularization improves stability when many quantiles are used (Cherodian et al., 23 Jan 2026).

7. Scope, Connections, and Extensions

Q-LS unifies a spectrum of robust estimation, regression, and distributional modeling techniques under the principle of quantile-matching via norm minimization:

  • It generalizes M-estimation, L1L_1-regression, least-trimmed/masked squares, and LqL_q-Wasserstein projection.
  • Recent algorithmic advances make robust Q-LS estimators computationally feasible for large-scale and high-dimensional data, with strong theoretical guarantees for small- and large-sample performance.
  • The choice of quantile levels, weighting, trimming, and constraint structure allows Q-LS estimators to target robustness, efficiency, parsimony, or tail accuracy as dictated by the application.

Q-LS estimators therefore provide flexible, scalable, and theoretically sound alternatives to classical likelihood-based procedures, especially when distributional robustness, heavy tails, or outlier resistance are essential (Adjieteh et al., 2024, Peng et al., 2023, Bertsimas et al., 2013, Puerto et al., 2024, Cherodian et al., 23 Jan 2026).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Quantile Least Squares (Q-LS).