Adversarial Robustness in Regression Models
- Adversarial robustness in regression is the study of models under worst-case, targeted perturbations to inputs and responses, emphasizing accuracy-robustness trade-offs and certified guarantees.
- It leverages minimax theory, convex regularization, and game-theoretic frameworks to balance standard risk with adversarial disruptions in parameter recovery and prediction quality.
- Practical approaches include adversarial min–max training, early stopping, and robust online algorithms that maintain performance in high-dimensional, overparameterized settings.
Adversarial robustness in regression refers to the theoretical, algorithmic, and empirical study of regression estimators and prediction models under worst-case, targeted perturbations to inputs, responses, or the training process. Unlike classical robustness that addresses random, possibly heavy-tailed or outlier contamination, adversarial robustness focuses on targeted, often norm-bounded or adaptive attacks calibrated to maximally disrupt prediction quality, recovery of ground-truth parameters, or fairness. The resulting theory spans minimax trade-offs, convex and nonconvex regularization, certified guarantees, PAC learnability, game-theoretic equilibrium, streaming algorithms, Bayesian inference, and modern deep neural network behavior.
1. Formal Definitions and Risk Frameworks
Let denote the feature and response; a regression model is specified by a predictor or parameter vector in linear models. The standard (population) risk is
with typically squared loss . The -adversarial risk under norm is
A central object of study is the trade-off curve (Pareto frontier) between and as a function of adversarial strength , model class, and structural constraints. The effect of adversarial training, min-max estimators, and adaptive online procedures is fundamental. In addition, threats may target inputs, responses, particular samples (poisoning), or fairness metrics.
2. Fundamental Accuracy–Robustness Trade-offs
Adversarial robustness in regression is governed by a suite of minimax lower bounds. A general abstract result (Bahmani, 2024) states: for suitable . In least-squares, with , this yields: where
is the mean local smoothness. Thus, any estimator attaining near-optimal standard risk must be “locally flat” (small ) in the adversarial norm, else its adversarial risk grows.
This trade-off is controlled by a Poincaré-type constant, encoding the alignment between the data distribution and the adversarial geometry: where is the covariance-induced norm, is the dual to the adversary’s norm. For high-dimensional -attacks, , making the trade-off especially stringent: non-trivial robustness is possible only for very small radii or low signal-to-noise ratio (Bahmani, 2024, Dohmatob et al., 2023).
This accuracy–robustness boundary manifests as explicit thresholds. For instance, for polynomial ridge regression, the necessary condition for robustness without sacrificing accuracy is
where denotes a signal-to-noise ratio for the regression problem (Bahmani, 2024).
3. Minimax, Algorithmic, and Statistical Characterizations
Linear Models and Explicit Solutions
For linear regression with -norm attacks and , adversarial risk decomposes as (Xing et al., 2020, Dohmatob et al., 2023): with . The minimizer is a ridge-regularized shrinkage: where solves an implicit equation depending on . This explicit link shows that adversarial robustness in linear regression naturally induces or regularization, depending on the attack norm (Ribeiro et al., 2022, Xie et al., 2024).
High-dimensional and Sparse Regimes
In sparse high-dimensional linear regression (), adversarial training under perturbations admits a convex dual form: and under restricted eigenvalue conditions, matches the classical minimax rates up to log factors (Xie et al., 2024, Ribeiro et al., 2022).
Group-structured adversarial training further narrows bounds when the signal exhibits group-sparsity. The group penalty,
yields reduced upper bounds whenever nonzeros cluster in few groups (Xie et al., 2024).
Overparameterization and Nonparametric Settings
Adversarial training in overparameterized regimes () causes sharp phase-transitions: for disturbances below a threshold , the solution jumps discontinuously to the minimum-norm interpolator (Ribeiro et al., 2022). In nonparametric regression, perfect interpolation can drastically worsen adversarial robustness, and even mild adversarial threats destroy the standard minimax rate if enforced (Peng et al., 22 Jan 2026).
4. Algorithms for Adversarially Robust Regression
Min–Max and Regularized Estimators
- Adversarial min–max training:
admits a dual convex form linking adversarial robustness to robust regression and parameter shrinkage (Ribeiro et al., 2022, Xie et al., 2024).
- Group adversarial training: convex objectives with group penalties enable adaptive control over both sparsity and group structure (Xie et al., 2024).
- Two-stage robustification: (1) estimate mean and covariance (by OLS, LASSO, thresholded estimators), (2) apply a risk-minimizing shrinkage determined by adversarial radius, yielding consistency and sharp minimax-optimal rates in both dense and sparse regimes (Xing et al., 2020).
Early Stopping and GD+ Approaches
- Gradient descent with early stopping can be near-minimax optimal for attacks but can be arbitrarily suboptimal for general Mahalanobis attacks. Feature-rescaling (GD+) schemes, using preconditioning by the square root of the adversary’s norm, restore close-to-optimality (Scetbon et al., 2023).
- Robust convex estimators: A two-stage approach based on the surrogate
achieves within a factor 1.11 of minimax risk for all norms, requiring only standard convex optimization tools.
Robust Streaming and Online Algorithms
- Online importance sampling (leveraged row sampling) supports adversarially robust streaming regression: at each arrival, the algorithm tosses fresh random bits to select rows via leverage scores; this approach is empirically and theoretically robust to adaptive, streaming adversaries, unlike fixed-projection sketching (Braverman et al., 2021).
- Spectral reweighting (SCRAM) and strong convex SDP relaxations provide distribution-free, online, and high-dimensional robust regression under Huber contamination, achieving the optimal dependence on the contamination fraction (Chen et al., 2020).
Fairness-aware and Causally Robust Methods
- Adversarially robust fair regression: Minimax frameworks incorporating mean-squared and group fairness under point or rank-one poisonings admit explicit solution strategies, maintaining both low accuracy loss and fairness under attack (Jin et al., 2022).
- Causal feature estimation via adversarial instrumental variable regression separates causal from non-causal (spurious) directions, and causal inoculation (CAFE) regularizes deep regression defenses by aligning features with their causal components (Kim et al., 2023).
Certified and PAC learning Perspectives
- PAC-Bayesian robust Bayesian regression: Generalized robust posteriors admit closed-form adversarial risk and nonvacuous generalization bounds (Sabanayagam et al., 20 Feb 2025).
- Certified randomized smoothing: Regression analogs of randomized smoothing provide explicit probabilistic certificates for bounded perturbations, based on concentration inequalities for Gaussian-perturbed average predictors (Rekavandi et al., 2024).
- Robust PAC learnability: Classes of bounded real-valued predictors have adversarially robust PAC learners if and only if the fat-shattering dimension is finite (Attias et al., 2022).
5. Empirical Phenomena and Phase Transitions
Phase-Transition Analysis
The accuracy–robustness trade-off often exhibits a phase-transition:
- For sufficiently small , robustification may be achieved at no cost in standard risk (“free-lunch” regime) (Dohmatob et al., 2023, Bahmani, 2024).
- Beyond critical thresholds, standard risk must deteriorate for additional gains in adversarial risk.
Overparametrized models (deep nets, minimum-norm solutions) are particularly fragile: strong interpolation may preserve generalization performance under classical risk but can destroy robustness to even minor adversarial input shifting. As the attack radius or the “influence” of interpolated points increases, adversarial risk can diverge logarithmically with sample size (Peng et al., 22 Jan 2026).
Practical Behavior
Key empirical takeaways from the literature:
- Min-max trained and group-penalized estimators empirically maintain minimax-optimal rates under attack, whereas classical estimators fail (Xie et al., 2024).
- Online leverage-based coreset methods preserve streaming regression accuracy under adversarial data adaptation, outperforming sketch-based methods, which can exhibit catastrophic failures once the adversary exploits the fixed sketch structure (Braverman et al., 2021).
- In deep networks, adversarial training and causal-feature regularization (CAFE) can provide gains in robustness with moderate cost to standard accuracy (Kim et al., 2023).
| Algorithmic approach | Certified against adaptive attacks? | Minimax-rate optimality |
|---|---|---|
| Importance-sampling streaming | Yes | Yes |
| Sketch-based streaming | No | No |
| Early stopping (GD) | Only for | Yes ( only) |
| Two-stage convex minimization | Yes (all norms) | Yes (all norms) |
6. Extensions: Structured, Fair, and Causal Robustness
Adversarial robustness increasingly incorporates additional desiderata beyond minimax risk:
- Fairness constraints under adversarial poisonings can be enforced via explicit minimax formulations combining prediction and group error gaps, admitting tractable global or saddle-point solutions (Jin et al., 2022).
- In multilearner or federated settings (multiple learners share data or are simultaneously attacked), the game-theoretic approach yields equilibrium models with higher robustness than standard regularization (Tong et al., 2018).
- For models exploiting latent group- or causal-structure in features, customized penalties or IV-type regularizations can improve worst-case error bounds (Xie et al., 2024, Kim et al., 2023).
7. Open Problems and Future Directions
Despite significant advances, several challenges remain:
- Precise characterization of the accuracy–robustness boundary for classes of nonlinear, nonconvex, or interpolating estimators and deep neural networks (Peng et al., 22 Jan 2026, Bahmani, 2024).
- Universal, data-agnostic algorithms reaching minimax adversarial risk without incurring additional sample complexity.
- Robust regression under arbitrary (possibly unbounded or structured) perturbation sets, e.g., certified randomized smoothing for general norms, or distributionally-robust objectives for OWL or group-structured regularizers.
- Integrating fairness, interpretability, and robust statistical estimation seamlessly into adversarially robust regression frameworks.
These directions represent a convergence of robust statistics, optimization, learning theory, and adversarial ML, as adversarial robustness in regression continues to be a vibrant and rapidly advancing field.