Papers
Topics
Authors
Recent
Search
2000 character limit reached

SmallML: Efficient Conformal Prediction

Updated 19 November 2025
  • SmallML is a framework for constructing prediction sets that guarantee exact finite-sample marginal coverage while minimizing set size.
  • It utilizes innovative methods like SAPS and adaptive thresholding to optimize nonconformity scores and improve model calibration.
  • Empirical results on benchmarks like ImageNet demonstrate significant reductions in average set sizes, enhancing efficiency over classical approaches.

SmallML (Small-Set Marginal Learning) refers to the suite of methods, theory, and algorithmic design principles surrounding minimization of conformal prediction set size (often called "efficiency") subject to exact finite-sample marginal coverage constraints. The central object of study is the prediction set C(x)C(x): given a pretrained classifier or regressor, and under a minimal exchangeability or i.i.d. assumption on calibration and test data, construct C(x)C(x) such that Pr{YC(X)}1α\Pr\{Y \in C(X)\} \ge 1-\alpha for a user-specified miscoverage rate α\alpha, while maintaining C(x)|C(x)| as small as possible on average. Recent advances, particularly in deep learning and large-scale benchmarks, have focused on refined score constructions, optimization of nonconformity functions, adaptive thresholding, and incorporation of model calibration and probabilistic structure—all towards achieving set sizes orders of magnitude smaller than those produced by prior methods.

1. Principles of Marginal Validity and Set Efficiency

The conformal prediction framework requires only the exchangeability of calibration and test data, making no assumptions about the accuracy or calibration of the underlying model. For a pre-trained model π^:XΔK1\hat{\pi}: \mathcal X \to \Delta^{K-1} (class probabilities), a held-out calibration set {(xi,yi)}i=1n\{(x_i, y_i)\}_{i=1}^n, and a user-chosen miscoverage level α\alpha, the classical workflow computes non-conformity scores S(xi,yi)S(x_i, y_i), sets a threshold τ\tau as the empirical C(x)C(x)0-quantile, and defines C(x)C(x)1 (Huang et al., 2023). By construction,

C(x)C(x)2

All subsequent developments—whether relying on softmax scores, instance difficulty, higher-order uncertainty, or distributional estimates—are constrained by this foundational guarantee. The challenge addressed by SmallML is to design C(x)C(x)3 and calibrate C(x)C(x)4 such that C(x)C(x)5 is as small as possible for typical C(x)C(x)6.

2. Sorted Adaptive Prediction Sets (SAPS) and Label Ranking

The Sorted Adaptive Prediction Sets (SAPS) framework illustrates a key innovation in SmallML: minimize dependence on miscalibrated probability tails by discarding all but the maximum predicted probability. For C(x)C(x)7-class deep classifiers, SAPS defines for each C(x)C(x)8 and random C(x)C(x)9: Pr{YC(X)}1α\Pr\{Y \in C(X)\} \ge 1-\alpha0 where Pr{YC(X)}1α\Pr\{Y \in C(X)\} \ge 1-\alpha1 is the rank of label Pr{YC(X)}1α\Pr\{Y \in C(X)\} \ge 1-\alpha2 in descending sorted predictions, and Pr{YC(X)}1α\Pr\{Y \in C(X)\} \ge 1-\alpha3 is a user-tuned hyperparameter (Huang et al., 2023). Only the maximum softmax value is retained; all other probabilistic information is replaced by Pr{YC(X)}1α\Pr\{Y \in C(X)\} \ge 1-\alpha4. The conformal set is determined by thresholding Pr{YC(X)}1α\Pr\{Y \in C(X)\} \ge 1-\alpha5 at the calibrated Pr{YC(X)}1α\Pr\{Y \in C(X)\} \ge 1-\alpha6. This design yields:

  • Average set sizes Pr{YC(X)}1α\Pr\{Y \in C(X)\} \ge 1-\alpha7 those of classical APS on ImageNet (APS: 20.95, SAPS: 2.98 at Pr{YC(X)}1α\Pr\{Y \in C(X)\} \ge 1-\alpha8).
  • Consistent marginal coverage at the prescribed Pr{YC(X)}1α\Pr\{Y \in C(X)\} \ge 1-\alpha9 level.
  • Smaller, instance-adaptive sets—"easy" examples with high confidence receive smaller α\alpha0, while "hard" instances expand accordingly.
  • Uniformly improved conditional coverage (ESCV) versus RAPS and APS.

Experiments on ImageNet, CIFAR-100, and CIFAR-10 consistently find SAPS to outperform alternatives in size and conditional error, with stable performance over reasonable calibration set sizes and hyperparameter choices.

3. Theoretical Guarantees and Coverage Properties

SmallML methods such as SAPS preserve exact marginal validity by standard conformal theory: if the set construction is "nested" (set size increases monotonically as the acceptance threshold grows laxer) and the threshold α\alpha1 is chosen as the empirical α\alpha2-quantile of α\alpha3, then for all distributions: [ \Pr[Y \in C_{1-\alpha}(X)] \geq

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to SmallML.