Papers
Topics
Authors
Recent
Search
2000 character limit reached

MAT: Mixed-Strategy Game of Adversarial Training in Fine-tuning

Published 27 Jun 2023 in cs.CL and cs.AI | (2306.15826v1)

Abstract: Fine-tuning large-scale pre-trained LLMs has been demonstrated effective for various NLP tasks. Previous studies have established that incorporating adversarial training during the fine-tuning stage can significantly enhance model generalization and robustness. However, from the perspective of game theory, such utilizations of adversarial training correspond to pure-strategy games, which are inherently limited in terms of the scope of their strategies, thereby still having room for improvement. In order to push the performance boundaries, we propose a novel Mixed-strategy Adversarial Training algorithm (MAT). Methodologically, we derive the Nash equilibrium of a mixed-strategy game for adversarial training using Entropy Mirror Descent to establish MAT by sampling method. To verify the effectiveness of MAT, we conducted extensive benchmark experiments on large-scale pre-trained models, such as BERT and RoBERTa. MAT significantly outperforms the state-of-the-art methods on both the GLUE and ANLI benchmarks in terms of generalization and robustness.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (52)
  1. Better fine-tuning by reducing representational collapse. In ICLR, 2021.
  2. Invariant risk minimization games. In ICML, 2020.
  3. Generalization and equilibrium in generative adversarial nets (gans). In ICML, 2017.
  4. The second PASCAL recognising textual entailment challenge. In Proceedings of the Second PASCAL Challenges Workshop on Recognising Textual Entailment, 2006.
  5. Mirror descent and nonlinear projected subgradient methods for convex optimization. Operations Research Letters, 2003.
  6. The fifth PASCAL recognizing textual entailment challenge. In Proceedings of the Second Text Analysis Conference, 2009.
  7. A large annotated corpus for learning natural language inference. In EMNLP, 2015.
  8. Language models are few-shot learners. In NeurIPS, 2020.
  9. SemEval-2017 task 1: Semantic textual similarity multilingual and crosslingual focused evaluation. In Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017), 2017.
  10. The PASCAL recognising textual entailment challenge. In First PASCAL Machine Learning Challenges Workshop, 2005.
  11. Training gans with optimism. In ICLR, 2018.
  12. BERT: pre-training of deep bidirectional transformers for language understanding. In NAACL-HLT, 2019.
  13. Automatically constructing a corpus of sentential paraphrases. In Proceedings of the Third International Workshop on Paraphrasing, 2005.
  14. The third PASCAL recognizing textual entailment challenge. In Proceedings of the ACL-PASCAL@ACL 2007 Workshop on Textual Entailment and Paraphrasing, 2007.
  15. Generative adversarial nets. In NeurIPS, 2014.
  16. Explaining and harnessing adversarial examples. In ICLR, 2015.
  17. Rmsprop: Divide the gradient by a running average of its recent magnitude. Coursera, 2012.
  18. Finding mixed nash equilibria of generative adversarial networks. In ICML, 2019.
  19. First quora dataset release: Question pairs. Technical report, Quora, 2017.
  20. SMART: robust and efficient fine-tuning for pre-trained natural language models through principled regularized optimization. In ACL, 2020.
  21. Is BERT really robust? A strong baseline for natural language attack on text classification and entailment. In AAAI, 2020.
  22. Adam: A method for stochastic optimization. In ICLR, 2015.
  23. Hector J. Levesque. The winograd schema challenge. In AAAI Spring Symposium, 2011.
  24. Datasets: A community library for natural language processing. In EMNLP, 2021.
  25. Preconditioned stochastic gradient langevin dynamics for deep neural networks. In AAAI, 2016.
  26. Deep text classification can be fooled. In IJCAI, 2018.
  27. Roberta: A robustly optimized BERT pretraining approach. CoRR, abs/1907.11692, 2019.
  28. Adversarial training for large neural language models. CoRR, abs/2004.08994, 2020.
  29. Towards deep learning models resistant to adversarial attacks. In ICLR, 2018.
  30. Adversarial training methods for semi-supervised text classification. In ICLR, 2017.
  31. John Nash. Non-cooperative games. Annals of Mathematics, 54(2):286–295, 1951.
  32. Arkadi Nemirovski and D. Yudin. Problem complexity and method efficiency in optimization. Wiley, 1983.
  33. Adversarial NLI: A new benchmark for natural language understanding. In ACL, 2020.
  34. Pytorch: An imperative style, high-performance deep learning library. In NeurIPS, 2019.
  35. Improving language understanding by generative pre-training. OpenAI blog, 2018.
  36. Language models are unsupervised multitask learners. OpenAI blog, 2019.
  37. Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research, 2020.
  38. Squad: 100, 000+ questions for machine comprehension of text. In EMNLP, 2016.
  39. Adversarial training for free! In NeurIPS, 2019.
  40. Recursive deep models for semantic compositionality over a sentiment treebank. In EMNLP, 2013.
  41. Intriguing properties of neural networks. In ICLR, 2014.
  42. FEVER: a large-scale dataset for fact extraction and verification. In NAACL-HLT, 2018.
  43. GLUE: A multi-task benchmark and analysis platform for natural language understanding. In ICLR, 2019.
  44. Transferable adversarial examples can efficiently fool topic models. Computers & Security, 118:102749, 2022.
  45. Neural network acceptability judgments. Transactions of the Association for Computational Linguistics, 2019.
  46. Bayesian learning via stochastic gradient langevin dynamics. In ICML, 2011.
  47. A broad-coverage challenge corpus for sentence understanding through inference. In NAACL-HLT, 2018.
  48. Transformers: State-of-the-art natural language processing. In EMNLP, 2020.
  49. You only propagate once: Accelerating adversarial training via maximal principle. In NeurIPS, 2019.
  50. Evalda: Efficient evasion attacks towards latent dirichlet allocation. In AAAI, 2021.
  51. Freelb: Enhanced adversarial training for natural language understanding. In ICLR, 2020.
  52. Adversarial regularization as stackelberg game: An unrolled optimization approach. In EMNLP, 2021.
Citations (2)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (3)

Collections

Sign up for free to add this paper to one or more collections.