Papers
Topics
Authors
Recent
Search
2000 character limit reached

Detecting Mode Collapse in Language Models via Narration

Published 6 Feb 2024 in cs.CL and cs.AI | (2402.04477v1)

Abstract: No two authors write alike. Personal flourishes invoked in written narratives, from lexicon to rhetorical devices, imply a particular author--what literary theorists label the implied or virtual author; distinct from the real author or narrator of a text. Early LLMs trained on unfiltered training sets drawn from a variety of discordant sources yielded incoherent personalities, problematic for conversational tasks but proving useful for sampling literature from multiple perspectives. Successes in alignment research in recent years have allowed researchers to impose subjectively consistent personae on LLMs via instruction tuning and reinforcement learning from human feedback (RLHF), but whether aligned models retain the ability to model an arbitrary virtual author has received little scrutiny. By studying 4,374 stories sampled from three OpenAI LLMs, we show successive versions of GPT-3 suffer from increasing degrees of "mode collapse" whereby overfitting the model during alignment constrains it from generalizing over authorship: models suffering from mode collapse become unable to assume a multiplicity of perspectives. Our method and results are significant for researchers seeking to employ LLMs in sociological simulations.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (33)
  1. Cognitive network science reveals bias in gpt-3, gpt-3.5 turbo, and gpt-4 mirroring math anxiety in high-school students. Big Data and Cognitive Computing, 7(33):124.
  2. Artificial intelligence and teachers’ new ethical obligations. The International Review of Information Ethics, 31(1).
  3. Jacob Andreas. 2022. Language models as agent models. In Findings of the Association for Computational Linguistics: EMNLP 2022, page 5769–5779, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
  4. Out of one, many: Using language models to simulate human samples. Political Analysis, page 1–15.
  5. Roland Barthes and Stephen Heath. 1977. Image, Music, Text: Essays, 13. [dr.] edition. Fontana, London.
  6. Latent dirichlet allocation. Journal of machine Learning research, 3(Jan):993–1022.
  7. Language models are few-shot learners. (arXiv:2005.14165). ArXiv:2005.14165 [cs].
  8. Seymour Benjamin Chatman. 1978. Story and Discourse: Narrative Structure in Fiction and Film. Cornell University Press, Ithaca, N.Y.
  9. Supervising strong learners by amplifying weak experts. (arXiv:1810.08575). ArXiv:1810.08575 [cs, stat].
  10. Bert: Pre-training of deep bidirectional transformers for language understanding.
  11. Stylometry with r: A package for computational text analysis. The R Journal, 8(1):107.
  12. Katherine Elkins and Jon Chun. 2020. Can gpt-3 pass a writer’s turing test? Journal of Cultural Analytics, 5(2).
  13. The pile: An 800gb dataset of diverse text for language modeling. (arXiv:2101.00027). ArXiv:2101.00027 [cs].
  14. Maarten Grootendorst. 2022. Bertopic: Neural topic modeling with a class-based tf-idf procedure. (arXiv:2203.05794). ArXiv:2203.05794 [cs].
  15. Studying the history of ideas using topic models. In Proceedings of the Conference on Empirical Methods in Natural Language Processing - EMNLP ’08, page 363, Honolulu, Hawaii. Association for Computational Linguistics.
  16. David I. Holmes. 1998. The evolution of stylometry in humanities scholarship. Literary and Linguistic Computing, 13(3):111–117.
  17. Minh Hua and Rita Raley. 2020. Playing with unicorns: Ai dungeon and citizen nlp.
  18. Mistral 7b. (arXiv:2310.06825). ArXiv:2310.06825 [cs].
  19. Scaling laws for neural language models. (arXiv:2001.08361). ArXiv:2001.08361 [cs, stat].
  20. Evaluation of mode collapse in generative adversarial networks.
  21. Edisa Lozić and Benjamin Štular. 2023. Fluent but not factual: A comparative analysis of chatgpt and other ai chatbots’ proficiency and originality in scientific writing for humanities. Future Internet, 15(1010):336.
  22. Sourcing pandemic news: A cross-national computational analysis of mainstream media coverage of covid-19 on facebook, twitter, and instagram. Digital Journalism, 9(9):1261–1285.
  23. OpenAI. 2023. Openai platform.
  24. Training language models to follow instructions with human feedback. (arXiv:2203.02155). ArXiv:2203.02155 [cs].
  25. Generative agents: Interactive simulacra of human behavior. (arXiv:2304.03442). ArXiv:2304.03442 [cs].
  26. Social simulacra: Creating populated prototypes for social computing systems. In Proceedings of the 35th Annual ACM Symposium on User Interface Software and Technology, UIST ’22, page 1–18, New York, NY, USA. Association for Computing Machinery.
  27. Language models are unsupervised multitask learners. page 24.
  28. Large language models can be easily distracted by irrelevant context. In Proceedings of the 40th International Conference on Machine Learning, volume 202 of ICML’23, page 31210–31227, Honolulu, Hawaii, USA. JMLR.org.
  29. Ben Swanson and Eugene Charniak. 2012. Native language detection with tree substitution grammars. In Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Short Papers - Volume 2, ACL ’12, page 193–197, USA. Association for Computational Linguistics.
  30. A report on the first native language identification shared task. In Proceedings of the Eighth Workshop on Innovative Use of NLP for Building Educational Applications, page 48–57, Atlanta, Georgia. Association for Computational Linguistics.
  31. Hoang Thanh-Tung and Truyen Tran. 2020. Catastrophic forgetting and mode collapse in gans. In 2020 International Joint Conference on Neural Networks (IJCNN), page 1–10.
  32. Llama 2: Open foundation and fine-tuned chat models. (arXiv:2307.09288). ArXiv:2307.09288 [cs].
  33. Twisty: A multilingual twitter stylometry corpus for gender and personality profiling. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC’16), page 1632–1637, Portorož, Slovenia. European Language Resources Association (ELRA).
Citations (6)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (1)

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.