Papers
Topics
Authors
Recent
Search
2000 character limit reached

The Unequal Opportunities of Large Language Models: Revealing Demographic Bias through Job Recommendations

Published 3 Aug 2023 in cs.CL, cs.AI, and cs.CY | (2308.02053v2)

Abstract: LLMs have seen widespread deployment in various real-world applications. Understanding these biases is crucial to comprehend the potential downstream consequences when using LLMs to make decisions, particularly for historically disadvantaged groups. In this work, we propose a simple method for analyzing and comparing demographic bias in LLMs, through the lens of job recommendations. We demonstrate the effectiveness of our method by measuring intersectional biases within ChatGPT and LLaMA, two cutting-edge LLMs. Our experiments primarily focus on uncovering gender identity and nationality bias; however, our method can be extended to examine biases associated with any intersection of demographic identities. We identify distinct biases in both models toward various demographic identities, such as both models consistently suggesting low-paying jobs for Mexican workers or preferring to recommend secretarial roles to women. Our study highlights the importance of measuring the bias of LLMs in downstream applications to understand the potential for harm and inequitable outcomes.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (33)
  1. Persistent Anti-Muslim Bias in Large Language Models. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society (Virtual Event, USA) (AIES ’21). Association for Computing Machinery, New York, NY, USA, 298–306. https://doi.org/10.1145/3461702.3462624
  2. Heather Antecol and Kelly Bedard. 2004. The racial wage gap: The importance of labor force attachment differences across black, Mexican, and white men. Journal of Human Resources 39, 2 (2004), 564–583.
  3. Bard 2023. Google AI Updates: Bard and New AI Features in Search. Retrieved May 7, 2023 from https://blog.google/technology/ai/bard-google-ai-search-updates/
  4. Language (Technology) is Power: A Critical Survey of “Bias” in NLP. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Online, 5454–5476. https://doi.org/10.18653/v1/2020.acl-main.485
  5. Stereotyping Norwegian Salmon: An Inventory of Pitfalls in Fairness Benchmark Datasets. In Annual Meeting of the Association for Computational Linguistics.
  6. Semantics derived automatically from language corpora contain human-like biases. Science 356, 6334 (2017), 183–186. https://doi.org/10.1126/science.aal4230 arXiv:https://www.science.org/doi/pdf/10.1126/science.aal4230
  7. ChatGPT 2023. Introducing ChatGPT. Retrieved May 7, 2023 from https://openai.com/blog/chatgpt
  8. Crawling The Internal Knowledge-Base of Language Models. In Findings of the Association for Computational Linguistics: EACL 2023. Association for Computational Linguistics, Dubrovnik, Croatia, 1856–1869. https://aclanthology.org/2023.findings-eacl.139
  9. Emilio Ferrara. 2023. Should ChatGPT be Biased? Challenges and Risks of Bias in Large Language Models. arXiv:2304.03738 [cs.CY]
  10. Maarten Grootendorst. 2022. BERTopic: Neural topic modeling with a class-based TF-IDF procedure. arXiv:2203.05794 [cs.CL]
  11. Online negative sentiment towards Mexicans and Hispanics and impact on mental well-being: A time-series analysis of social media data during the 2016 United States presidential election. Heliyon 6, 9 (2020).
  12. HuggingChat 2023. HuggingChat. Retrieved May 7, 2023 from https://huggingface.co/chat/
  13. HuggingFace 2022. sentence-transformers/all-MiniLM-L6-v2. Retrieved May 7, 2023 from https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2
  14. Bias Out-of-the-Box: An Empirical Analysis of Intersectional Occupational Biases in Popular Generative Language Models. arXiv:2102.04130 [cs.CL]
  15. Li Lucy and David Bamman. 2021. Gender and Representation Bias in GPT-3 Generated Stories. In Proceedings of the Third Workshop on Narrative Understanding. Association for Computational Linguistics, Virtual, 48–55. https://doi.org/10.18653/v1/2021.nuse-1.5
  16. A Holistic Approach to Undesired Content Detection in the Real World. arXiv:2208.03274 [cs.CL]
  17. Douglas S Massey. 2009. Racial formation in theory and practice: The case of Mexicans in the United States. Race and social problems 1 (2009), 12–26.
  18. Robert W. McGee. 2023. Is Chat Gpt Biased Against Conservatives? An Empirical Study. (15 February 2023). https://doi.org/10.2139/ssrn.4359405
  19. UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction. arXiv:1802.03426 [stat.ML]
  20. A Survey on Bias and Fairness in Machine Learning. ACM Comput. Surv. 54, 6, Article 115 (jul 2021), 35 pages. https://doi.org/10.1145/3457607
  21. StereoSet: Measuring stereotypical bias in pretrained language models. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Association for Computational Linguistics, Online, 5356–5371. https://doi.org/10.18653/v1/2021.acl-long.416
  22. Dissecting racial bias in an algorithm used to manage the health of populations. Science 366, 6464 (2019), 447–453. https://doi.org/10.1126/science.aax2342 arXiv:https://www.science.org/doi/pdf/10.1126/science.aax2342
  23. Orestis Papakyriakopoulos and Ethan Zuckerman. 2021. The media during the rise of trump: Identity politics, immigration,” Mexican” demonization and hate-crime. In Proceedings of the International AAAI Conference on Web and Social Media, Vol. 15. 467–478.
  24. On Natural Language User Profiles for Transparent and Scrutable Recommendation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’22).
  25. Cordelia W. Reimers. 1983. Labor Market Discrimination Against Hispanic and Black Men. The Review of Economics and Statistics 65, 4 (1983), 570–579. http://www.jstor.org/stable/1935925
  26. David Rozado. 2023. The Political Biases of ChatGPT. Social Sciences 12, 3 (2023). https://doi.org/10.3390/socsci12030148
  27. The Self-Perception and Political Biases of ChatGPT. arXiv:2304.07333 [cs.CY]
  28. Quantifying Social Biases Using Templates is Unreliable. arXiv:2210.04337 [cs.CL]
  29. The Woman Worked as a Babysitter: On Biases in Language Generation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Association for Computational Linguistics, Hong Kong, China, 3407–3412. https://doi.org/10.18653/v1/D19-1339
  30. LLaMA: Open and Efficient Foundation Language Models. arXiv:2302.13971 [cs.CL]
  31. Investigating Gender Bias in Language Models Using Causal Mediation Analysis. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (Eds.), Vol. 33. Curran Associates, Inc., 12388–12401. https://proceedings.neurips.cc/paper_files/paper/2020/file/92650b2e92217715fe312e6fa7b90d82-Paper.pdf
  32. A prompt pattern catalog to enhance prompt engineering with chatgpt. arXiv preprint arXiv:2302.11382 (2023).
  33. Exploring AI Ethics of ChatGPT: A Diagnostic Analysis. arXiv:2301.12867 [cs.CL]
Citations (25)

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.