DP-EBMs: Private, Explainable Boosting
- DP-EBMs are interpretable GAMs that blend boosting methods with Gaussian noise injection to enforce (ε,δ)-differential privacy on sensitive tabular data.
- They integrate additive models with privacy-preserving techniques like residual clipping and budget partitioning across boosting rounds to balance accuracy and privacy.
- DP-EBMs offer both global and local interpretability, allowing experts to visualize and edit feature contributions in real-world applications such as healthcare and finance.
Differentially-Private Explainable Boosting Machines (DP-EBMs) are a class of interpretable statistical learning models that combine the additive transparency of Generalized Additive Models (GAMs) with rigorous (ε,δ)-differential privacy guarantees. DP-EBMs have emerged as state-of-the-art models for learning from sensitive tabular data in settings such as healthcare, finance, and social science, offering both high predictive accuracy and global/local interpretability. These models implement differential privacy by injecting Gaussian noise into carefully chosen algorithmic primitives during boosting-based GAM training. Rigorous analysis demonstrates that DP-EBMs provide strong privacy protection while incurring minimal loss in accuracy and preserving the ability to explain and edit learned models.
1. Model Definition and Architecture
Explainable Boosting Machines are specialized GAMs with the canonical form:
where denotes the conditional response, is a link function (identity for regression, log-odds for binary classification), is an intercept, and are univariate shape functions over each feature learned via sequential boosting (Nori et al., 2021). Each boosting round selects a feature, computes residuals, and fits a shallow tree (usually a single split, i.e., a decision stump), partitioning the feature into histogram bins and updating accordingly.
By construction, EBMs maintain strict additivity, allowing exact interpretation of both global effects (by plotting ) and local contributions (by evaluating per example). Their practical value is underscored in domains requiring both interpretability and robust statistical guarantees on data usage.
2. Incorporating Differential Privacy: Noise Injection Mechanism
DP-EBMs ensure (ε,δ)-differential privacy by perturbing the sufficient statistics used to update each at every boosting iteration with Gaussian noise, following the Gaussian Mechanism (Niu et al., 2022, Nori et al., 2021). At boosting round and feature , binned residual sums
are first subject to clipping to tightly bound sensitivity. Then, for each bin ℓ, Gaussian noise
with
is added. The privacy budget (ε,δ) is partitioned evenly across all boosting rounds, i.e., , . The resulting privatized sufficient statistics are then used to update the functions.
The model does not privatize the choice of splitting points; privacy is spent only on the computed values, not on tree structure or bin assignments. Empirically, this approach gives very tight accuracy bounds and avoids compounding privacy loss over combinatorial choices (Nori et al., 2021).
3. DP-EBM Training Algorithm
A high-level version of the differentially-private training procedure is:
- Initialization: Set and for all ; select number of rounds , clipping bound , and learning rate .
- Privacy Budgeting: Allocate , per round.
- Boosting Loop ( to ):
- (a) Compute residuals .
- (b) Clip each to .
- (c) Select feature .
- (d) Partition into histogram bins.
- (e) For each bin ℓ, compute for members, then add Gaussian noise to release .
- (f) Update the shape function via (where is the bin count).
- (g) Update as appropriate.
- (h) Retain other unchanged if .
- Output: .
For classification, residuals are replaced with gradients of the log-loss function. Binning of continuous features is performed with differentially-private quantile binning, consuming a small portion of the total budget (Nori et al., 2021).
4. Privacy Guarantees and Analytical Composition
Each boosting round is an application of the Gaussian Mechanism, achieving -DP per round based on the advanced composition theorem [Dwork–Roth]. Over all rounds, DP-EBM composes these guarantees to achieve overall -DP (Niu et al., 2022):
is sufficient for all updates, ensuring the desired global guarantee.
More refined analysis employs Gaussian Differential Privacy (GDP), leveraging tighter composition bounds [Dong et al. ’19]. Under this framework, each iteration corresponds to -GDP, and the cumulative effect is expressed as
where is the number of epochs and is the number of features (Nori et al., 2021). GDP-to- translation is then used for final accounting.
No further data access or post-hoc fitting is performed after the noisy sums, satisfying post-processing invariance.
5. Interpretability and Post-Training Model Editing
DP-EBMs provide exact global and local interpretability, as each can be directly plotted and interrogated. This transparency is preserved regardless of the injected noise (Nori et al., 2021). The post-processing property of DP allows for post-training editing—such as monotonicity enforcement via isotonic regression or manual smoothing of noisy shape functions—without any additional privacy cost. This is relevant for applications where expert correction of spurious artifacts (e.g., non-monotonicities induced by noise) is necessary prior to model deployment.
6. Empirical Evaluation and Statistical Trade-offs
Empirical results demonstrate that DP-EBMs maintain high accuracy even under strong privacy regimes, significantly outperforming prior art such as DPBoost and private linear/logistic regression on standard tabular datasets (Nori et al., 2021). Illustrative results on the Adult Income dataset (AUROC):
| ε | DPBoost | DP Logistic | DP-EBM (classic) | DP-EBM (GDP) | Non-private EBM |
|---|---|---|---|---|---|
| 0.5 | 0.558 | 0.488 | 0.873 ± 0.007 | 0.875 ± 0.005 | 0.923 ± 0.003 |
| 1.0 | 0.566 | 0.471 | 0.880 ± 0.006 | 0.883 ± 0.005 | 0.923 ± 0.003 |
| 4.0 | 0.734 | 0.549 | 0.889 ± 0.004 | 0.889 ± 0.004 | 0.923 ± 0.003 |
For regression, DP-EBMs remain close to non-private RMSE and outperform alternatives at all privacy levels.
On mean squared error (MSE) decomposition for doubly-robust CATE estimation, variance increases markedly as (≈10×), with bias increasing only modestly (≈2×) (Niu et al., 2022). Representative table on a voting-turnout dataset ():
| ε | MSE | Bias² | Var |
|---|---|---|---|
| 16 | 0.018 | 0.012 | 0.006 |
| 8 | 0.022 | 0.014 | 0.008 |
| 4 | 0.035 | 0.017 | 0.018 |
| 2 | 0.064 | 0.020 | 0.044 |
| 1 | 0.120 | 0.025 | 0.095 |
This suggests that most privacy-induced accuracy loss arises from increased output variance rather than estimator bias.
Interpretability degrades with stronger privacy (small ε) as learned become less smooth and more "jumpy", but the directional, qualitative ordering of feature effects is generally maintained (Niu et al., 2022).
7. Practical Applications and Implementation
DP-EBMs have been deployed in domains where both privacy and interpretability are paramount, including healthcare, criminal justice, and finance (Nori et al., 2021). The methodology is implemented and distributed in the open-source InterpretML Python library, with automated handling of privacy budget allocation, per-round clipping, Gaussian noise addition, and composition accounting.
No statistical trade-off is incurred for post-hoc editing, ensuring that expert-in-the-loop corrections for domain alignment do not degrade privacy guarantees. Models remain fully auditable and explainable after privatized training, with no additional data access required.
References:
- "Differentially Private Estimation of Heterogeneous Causal Effects" (Niu et al., 2022)
- "Accuracy, Interpretability, and Differential Privacy via Explainable Boosting" (Nori et al., 2021)