Papers
Topics
Authors
Recent
Search
2000 character limit reached

Distributionally and Adversarially Robust Logistic Regression via Intersecting Wasserstein Balls

Published 18 Jul 2024 in math.OC and cs.LG | (2407.13625v4)

Abstract: Adversarially robust optimization (ARO) has emerged as the de facto standard for training models that hedge against adversarial attacks in the test stage. While these models are robust against adversarial attacks, they tend to suffer severely from overfitting. To address this issue, some successful methods replace the empirical distribution in the training stage with alternatives including (i) a worst-case distribution residing in an ambiguity set, resulting in a distributionally robust (DR) counterpart of ARO; (ii) a mixture of the empirical distribution with a distribution induced by an auxiliary (e.g., synthetic, external, out-of-domain) dataset. Inspired by the former, we study the Wasserstein DR counterpart of ARO for logistic regression and show it admits a tractable convex optimization reformulation. Adopting the latter setting, we revise the DR approach by intersecting its ambiguity set with another ambiguity set built using the auxiliary dataset, which offers a significant improvement whenever the Wasserstein distance between the data generating and auxiliary distributions can be estimated. We study the underlying optimization problem, develop efficient solution algorithms, and demonstrate that the proposed method outperforms benchmark approaches on standard datasets.

Summary

  • The paper introduces a novel framework that combines distributionally robust and adversarially robust optimization for logistic regression.
  • It leverages intersecting Wasserstein balls centered on empirical and auxiliary datasets to mitigate overfitting and improve out-of-sample performance.
  • Empirical evaluations on UCI and MNIST datasets demonstrate significant robustness gains compared to classical methods.

Distributionally and Adversarially Robust Logistic Regression via Intersecting Wasserstein Balls

Introduction

The paper explores a novel approach to robust logistic regression, addressing both distributional uncertainty and adversarial attacks by utilizing intersecting Wasserstein balls. The failure of empirical risk minimization (ERM) to withstand adversarial attacks necessitates a method ensuring robustness against such perturbations. Traditional adversarially robust optimization (ARO) models, while providing adversarial defense, are susceptible to severe overfitting. The authors propose a framework that combines the strengths of distributionally robust optimization (DRO) with ARO, using auxiliary datasets to refine the decision boundary of logistic regression. This method offers enhanced out-of-sample performance, managing the trade-off between robustness and overfitting. Figure 1

Figure 1

Figure 1: Out-of-sample errors under varying attack strengths (left) and runtimes under varying numbers of empirical and auxiliary instances (right) of artificial experiments.

Methodology

Wasserstein DRO Framework:

The proposed approach formulates a distributionally robust counterpart of ARO. The method involves intersecting the traditional Wasserstein ball, centered on the empirical distribution, with another centered on an auxiliary distribution. The intersection mitigates model conservatism and enhances adaptability to true data distribution shifts. Auxiliary datasets could be synthetic or out-of-domain, offering a refined ambiguity set to counteract adversarial effects effectively. This technique capitalizes on the tractability of the Wasserstein distance and its application in ML robustness. Figure 2

Figure 2: Runtimes under a varying number of features in the artificially generated empirical and auxiliary datasets.

Algorithmic Solution:

The authors reformulate robust logistic regression as a convex optimization problem. An efficient approximation algorithm is introduced, leveraging dual formulations to handle the complexities of intersecting ambiguity sets. The solution maintains a balance among computational efficiency, robust defense, and minimizing out-of-sample error.

Experimental Results

Extensive experiments on UCI datasets and MNIST/EMNIST datasets validate the practical efficacy of the proposed method. The results consistently demonstrate superior out-of-sample performance compared to both classical ARO and ERM methods. The Wasserstein DRO approach, particularly when enriched with auxiliary data, shows significant improvements in robustness without compromising model performance on unperturbed data.

Implications

This research provides a significant leap in robust statistics and ML, accommodating adversarial examples' increasing sophistication. The ability to intersect multiple Wasserstein balls introduces a new paradigm in uncertainty quantification, emphasizing the importance of auxiliary datasets in training robust models.

Conclusion

The intersection of Wasserstein balls offers a powerful mechanism to enhance logistic regression's adversarial and distributional robustness. By addressing both adversarial perturbations and distributional shifts, the proposed framework strikes an ideal balance, offering robust, scalable solutions suitable for various applications. Future research could explore extensions to other loss functions and address computational challenges in high dimensional settings.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 4 likes about this paper.