Evaluating Large Language Models through Gender and Racial Stereotypes
Abstract: LLMs have ushered a new age of AI gaining traction within the NLP community as well as amongst the general population. AI's ability to make predictions, generations and its applications in sensitive decision-making scenarios, makes it even more important to study these models for possible biases that may exist and that can be exaggerated. We conduct a quality comparative study and establish a framework to evaluate LLMs under the premise of two kinds of biases: gender and race, in a professional setting. We find out that while gender bias has reduced immensely in newer models, as compared to older ones, racial bias still exists.
- Unmasking Contextual Stereotypes: Measuring and Mitigating BERT’s Gender Bias. CoRR abs/2010.14534 (2020). arXiv:2010.14534 https://arxiv.org/abs/2010.14534
- Language Models are Few-Shot Learners. arXiv:2005.14165Â [cs.CL]
- Semantics derived automatically from language corpora contain human-like biases. Science 356, 6334 (14 April 2017), 183–186. https://doi.org/10.1126/science.aal4230
- Scaling Instruction-Finetuned Language Models. arXiv:2210.11416Â [cs.LG]
- Devah and Shepherd. 1997. Racial, societal and class bias in clinical judgement. arXiv:https://doi.org/10.1111/j.1468-2850.1997.tb00104.x
- BERT: Pre-training of deep bidirectional transformers for language understanding. In NAACL HLT 2019 - 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies - Proceedings of the Conference, Vol. 1. 4171–4186. www.scopus.com Cited By :6289.
- RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv:1907.11692Â [cs.CL]
- Science faculty’s subtle gender biases favor male students. Proceedings of the National Academy of Sciences 109, 41 (2012), 16474–16479. https://doi.org/10.1073/pnas.1211286109 arXiv:https://www.pnas.org/doi/pdf/10.1073/pnas.1211286109
- Muhammad Ali Pervez. 2010. Impact of Emotions on Employee’s Job Performance: An Evidence from Organizations of Pakistan. arXiv:https://ssrn.com/abstract=1668170
- Language Models are Unsupervised Multitask Learners.
- Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. arXiv:1910.10683Â [cs.LG]
- Exploring Perception Of Professionals Regarding Introversion And Extroversion In Relation To Success At Workplace. 7 (01 2021).
- Hanvold T.N. Sterud T. 2021. Effects of adverse social behaviour at the workplace on subsequent mental distress: a 3-year prospective study of the general working population in Norway. arXiv:https://doi.org/10.1007/s00420-020-01581-y
- LaMDA: Language Models for Dialog Applications. arXiv:2201.08239Â [cs.CL]
- XLNet: Generalized Autoregressive Pretraining for Language Understanding. arXiv:1906.08237Â [cs.CL]
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.