Papers
Topics
Authors
Recent
Search
2000 character limit reached

Adaptive Minibatch Subsampling in Deep Learning

Updated 16 January 2026
  • The paper introduces UIDS, a framework using first- and second-order influence functions to guide adaptive, unweighted subsampling, improving risk reduction and generalization.
  • The methodology combines greedy batch selection and probabilistic influence sampling to efficiently compute Hessian and gradient metrics for data point selection.
  • Empirical results demonstrate that UIDS achieves faster convergence and lower error rates compared to traditional subsampling methods in classification and regression tasks.

Unweighted Influence Data Subsampling (UIDS) is a principled algorithmic framework that leverages statistical influence functions to select subsets of data for model training, aiming to maximize statistical efficiency and predictive performance under unweighted subsample regimes. Unlike classical weighted subsampling schemes, UIDS constructs subsamples in which all retained data points contribute equally (i.e., unit weight) during model fitting, while selection is guided by influence scores derived from first-order perturbations of the empirical risk minimizer or related objectives. This approach yields improved risk reduction, computational efficiency, and, under certain regimes, can achieve out-of-sample generalization superior to models trained on the full dataset (Raj et al., 2020, Wang et al., 2019, Ting et al., 2017).

1. Formalism and Influence Function Theory

UIDS is predicated on the use of influence functions, which quantify the infinitesimal impact of perturbing or reweighting a given data point on the model parameter θ^\hat\theta estimated by empirical risk minimization. For a training set Z={zi}i=1NZ=\{z_i\}_{i=1}^N and loss function (z,θ)\ell(z,\theta), the estimator is

θ^(Z)=argminθΘi=1N(zi,θ).\hat\theta(Z) = \arg\min_{\theta \in \Theta} \sum_{i=1}^N \ell(z_i, \theta).

The (Bouligand) influence function for zz is

BIF(z;θ^,Z)=HZ1θ(z,θ^(Z)),\mathrm{BIF}(z; \hat\theta, Z) = - H_Z^{-1} \nabla_\theta \ell(z, \hat\theta(Z)),

with HZ=i=1Nθ2(zi,θ^(Z))H_Z = \sum_{i=1}^N \nabla^2_\theta \ell(z_i, \hat\theta(Z)) the empirical Hessian, assumed invertible under standard regularity (Raj et al., 2020).

Second-order influence functions extend to the effect of upweighting a sample on the out-of-sample risk R(θ^)R(\hat\theta), especially relevant for model selection and validation. For test point zjz_j,

ϕ(zi,zj)=(θ(θ^;zj))Hθ^1θ(θ^;zi),\phi(z_i, z_j) = - (\nabla_\theta \ell(\hat\theta; z_j))^\top H^{-1}_{\hat\theta} \nabla_\theta \ell(\hat\theta; z_i),

and the total influence of ziz_i over a test/validation set VV is

ϕi=1VjVϕ(zi,zj).\phi_i = \frac{1}{|V|} \sum_{j \in V} \phi(z_i, z_j).

This formalism enables model-specific and risk-sensitive subsampling (Wang et al., 2019).

2. UIDS Algorithmic Schemes

Greedy Batch Influence Selection

A general instance of UIDS for subset selection proceeds iteratively. At each step, it fits parameters θ^S\hat\theta_S on the current subset SS, computes candidate influence scores ΔR(z)\Delta_R(z) for zZSz \in Z \setminus S (projected reduction in risk RR), then adds the top mm points with the maximal positive ΔR(z)\Delta_R(z): ΔR(z)=θR(θ^S)HS1θ(z,θ^S).\Delta_R(z) = -\nabla_\theta R(\hat\theta_S)^\top H_S^{-1} \nabla_\theta \ell(z, \hat\theta_S). For validation risk, with VV the validation set: Δval(z)=(vVθ(v,θ^S))HS1θ(z,θ^S).\Delta_{\mathrm{val}}(z) = - \left( \sum_{v\in V} \nabla_\theta \ell(v, \hat\theta_S) \right)^\top H_S^{-1} \nabla_\theta \ell(z, \hat\theta_S). An ε\varepsilon-greedy variant can mix top-influence and random points to avoid overfitting (Raj et al., 2020).

Probabilistic Influence Sampling

UIDS may also operate as a probabilistic scheme, replacing hard dropout (select all with ϕi<0\phi_i < 0) by sampling probability πi=π(ϕi)\pi_i = \pi(\phi_i) decreasing in total influence:

  • Linear: πi=clip(1αϕi,0,1)\pi_i = \mathrm{clip}(1 - \alpha \cdot \phi_i, 0, 1)
  • Sigmoid: πi=1/(1+exp(αϕi/Δ))\pi_i = 1 / (1 + \exp(\alpha \cdot \phi_i / \Delta)) Points are drawn iid Bernoulli(πi\pi_i), resulting in subsets whose risk behavior is controlled by the smoothness of π\pi (Wang et al., 2019).

Unweighted Influence Estimation

In the variant from (Ting et al., 2017), for estimators admitting asymptotic linear influence function expansions: θ^(PN)=θ(P)+1Ni=1Nψ(Xi;θ)+op(1),\hat\theta(P_N) = \theta(P) + \frac{1}{N} \sum_{i=1}^N \psi(X_i; \theta) + o_p(1), the optimal Poisson sampling probability is piψip_i \propto \|\psi_i\|, scaled to the chosen subsample size nn. Then, forming pseudo-observations ui=ψi/piu_i = \psi_i / p_i,

θUIDS=θ0+1niSui,\theta_\mathrm{UIDS} = \theta_0 + \frac{1}{n} \sum_{i \in S} u_i,

where θ0\theta_0 is a pilot estimate. Each sampled case is assigned unit weight for retraining.

3. Theoretical Guarantees

UIDS algorithms are accompanied by several nonasymptotic and asymptotic performance bounds:

  • Risk Reduction: For greedy batch selection, after pp adaptive steps, the expected risk of UIDS subset θM+pmg\theta_{M+pm}^g is

R(θM+pmg)E[R(θM+pmr)]i=1pAM+(i1)m+Gm2i=1p1(M+im)2,R(\theta_{M+pm}^g) - \mathbb{E}[R(\theta_{M+pm}^r)] \le \sum_{i=1}^{p} A'_{M+(i-1)m} + G\, m^2 \sum_{i=1}^p \frac{1}{(M + i m)^2},

where AA' is the instantaneous gain over random and GG a constant. This guarantees UIDS is strictly better than uniform if some points have large influence (Raj et al., 2020).

  • Subset Superiority: If the covariance between perturbations and influence, Cov(ϕ,ϵ)\mathrm{Cov}(\phi, \epsilon), is negative, then the subset model risk improves over the full set: R(θ^ϵ;Q)R(θ^;Q)R(\hat\theta_\epsilon; Q') \le R(\hat\theta; Q'). Lemma 1 shows the sample mean of influence is zero, so subsets chosen with negative correlation to ϕ\phi are systematically superior (Wang et al., 2019).
  • Distributional Robustness: For probabilistic ϵ(ϕ)\epsilon(\phi) maps with bounded gradient, worst-case risk over a χ2\chi^2-ball of nearby distributions is Lipschitz in ϕ\phi, enabling regulation of generalization behavior and avoidance of overfitting on a validation set (Wang et al., 2019).
  • Variance Optimality: UIDS achieves the minimal possible asymptotic variance for estimating θ^\hat\theta via subsampling, matching the Horvitz–Thompson estimator while enabling unweighted fitting (Ting et al., 2017).

4. Computational Complexity and Practical Considerations

UIDS implementations are dominated by linear algebraic operations on the Hessian HH and gradient vectors.

  • Hessian Operations: Naïve inversion scales as O(d3)O(d^3) (dimension dd), but conjugate gradient or sketching can reduce this to O(d2)O(d^2) per Hessian-vector product (Raj et al., 2020).
  • Influence Vector Computation: For nn points and kk iterations in preconditioned conjugate gradient (PCG), costs approximate O(ndk+nd)O(n d k + n d) (Wang et al., 2019).
  • Memory Footprint: Storing HH, ss, and ϕ\phi costs O(d2)O(d^2) and O(n)O(n).
  • Retraining: Final ERM on the selected subset scales as O(rndT)O(r n d T'), where r<1r < 1 is the sampling ratio.

UIDS is computationally tractable for GLMs and moderate to large nn when using efficient linear solvers. Approximate Hessian methods (sketching, Hessian-free CG) scale to datasets with 10610^610810^8 samples. Batch size (mm) and seed size (MM) require practical tuning; small mm controls greedy approximation error, and MdM \approx d ensures invertibility.

UIDS generalizes to non-differentiable models (e.g. tree ensembles) via proxy linear models, with strong empirical results on model transfer (Raj et al., 2020).

5. Empirical Outcomes and Benchmarks

UIDS has been evaluated on diverse tasks:

  • Model Selection: On Amazon employee-access and MNIST datasets, UIDS reaches target accuracy with 40–50% fewer samples than uniform subsampling (Raj et al., 2020).
  • Regression: California housing regression achieves RMSE 1.5 (UIDS) vs. 2.0 (random) for same subset sizes.
  • Hyperparameter Tuning: For random forest + Hyperband on Boston housing, UIDS finds optimal configurations 2× faster than random baseline.
  • Large-Scale Classification: On text, image, and CTR datasets (UCI, Criteo, Avazu, industrial sets), UIDS (linear/sigmoid variants) attains lower out-of-sample log-loss than uniform, weighted, or dropout schemes. On the 100M-sample “Company” dataset, UIDS achieves test log-loss of 0.1952 versus 0.1955 for the full set (Wang et al., 2019).
  • Subset-Model Superiority: Empirical results confirm that subset models via UIDS can outperform full-set ERM on future risk (Wang et al., 2019).

6. Limitations and Practical Guidance

UIDS efficacy depends on model and data regularity:

  • Hessian Nondegeneracy: Non-convex losses (e.g. deep networks) may violate invertibility, degrading influence estimates.
  • Influence Variability: Substantial gains over random sampling appear when data exhibits heterogeneous influence.
  • Overfitting: Pure dropout strategies can overfit validation sets, yielding poor generalization on unseen data; smoothing the sampling map (π) mitigates this risk (Wang et al., 2019).
  • Subset Size: Very small subsets or extreme dimensionality (dnd \gg n) can introduce instability; initial subset size MdM \approx d is recommended.
  • Memory Constraints: For very large dd, memory and computational demands call for approximate methods.
  • Applicability: For non-differentiable models, proxies enable effective transfer but may not fully match target model behavior.

Guidance suggests practical batch sizes m[1,10]m \in [1,10], validation set size 10–20% for influence scoring, and ε\varepsilon-greedy random mixes (ε0.1\varepsilon \sim 0.1) for robustness.

UIDS distinguishes itself from weighted influence sampling and leverage-score or gradient-based subsampling:

  • Weighted Schemes: Horvitz–Thompson weighting achieves minimal variance but requires specialized handling for weights; UIDS attains equivalent variance with unweighted retraining (Ting et al., 2017).
  • Leverage and Gradient Sampling: Solely exploiting leverage scores (design-matrix geometry) or gradients is less efficient than influence-based weighting, especially when both sources of variability are important.
  • Dropout Subsampling: Data dropout (drop all ϕi>0\phi_i > 0) is brittle; UIDS's smooth probabilistic sampling achieves better robustness and generalization (Wang et al., 2019).
  • Computational Tradeoff: Weighted influence sampling may be computationally more intensive due to repeated influence calculation; UIDS mitigates this by single-pass calculation and unweighted fitting.

Empirical comparisons demonstrate UIDS requires 30–50% as many samples as uniform subsampling to achieve comparable estimator error in regression tasks, and consistently outperforms leverage-only and gradient-only schemes.


UIDS, as formalized in multiple frameworks, leverages influence-function theory to deliver flexible, statistically efficient, and computationally tractable subsampling algorithms for modern machine learning and statistical estimation tasks (Raj et al., 2020, Wang et al., 2019, Ting et al., 2017).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (3)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Adaptive Minibatch Subsampling in Deep Learning.