Examining Gender Bias in Language Models and Humans
A study reveals parallels in gender bias between language models and human decision-making.
― 6 min read
Table of Contents
In recent years, researchers have discovered that language models often show biases similar to those found in people. This is especially true when it comes to gender-related biases that affect how pronouns refer to people in sentences. The focus of this research is to understand how these biases in models compare to those in humans, particularly in the context of Coreference Resolution, where a pronoun must correctly refer back to a noun in a sentence.
Gender Bias in language use can arise from societal norms, leading both humans and language models to make quick judgments based on these norms. Humans, when reading, can get influenced by stereotypes, which can result in biased interpretations of sentences. Similarly, models trained on large datasets can pick up these biases and rely on them to make predictions.
The central question this research addresses is whether the biases seen in language models reflect human behavior. To explore this, researchers turn to the dual-process theory, a concept from psychology that describes two systems of thinking. The first system is fast, automatic, and often leads to biases without much thought. The second system is slower, more thoughtful, and can correct initial judgments if there is time to reflect.
To investigate these ideas, researchers conducted two experiments involving human participants. In the first experiment, they used self-paced reading tasks. Participants read sentences where they had to quickly decide which pronoun referred to which noun, a process that mimics natural reading. This method allows researchers to understand the quick, automatic responses people make.
In the second experiment, participants answered questions after reading sentences. This task required more conscious thought and gave insights into the slower, more deliberate decision-making process. Researchers examined how time constraints affected participants’ responses, revealing more about the influence of quick judgments in biases.
From these experiments, it was found that humans made slightly more biased decisions than the models when analyzing real-world sentences. However, with synthetic sentences, which were more controlled and less varied, models showed greater bias. This disparity raises questions about how different types of sentences affect bias.
The researchers categorized biases into two main areas. On one side are annotation artifacts, which are biases that exist only in specific training datasets and don’t reflect real-world language use. On the other side are human-like biases, which may help in some contexts but can also lead to harmful outcomes.
To analyze these biases further, researchers created interfaces for human annotations, allowing them to better compare the performance of models and humans. Specifically, they looked at how gender bias manifests in coreference resolution tasks in English. They discovered that humans often lean towards stereotypical interpretations of sentences, which can lead to biased conclusions.
The study focused on three datasets designed to identify gender bias in coreference resolution. These included both synthetic data, made up of sentences with a specific structure, and more natural data gathered from real-world sources. The synthetic data allowed for controlled comparisons, while the natural data offered a more accurate reflection of how people read and interpret language.
By using a method called the MAZE task, which requires participants to choose the next word in a sentence from two options, researchers aimed to understand the timing of reading decisions. This incremental processing method provides insights into how quickly and effectively people can resolve pronouns based on previous context.
Key Findings from Experiments
The experiments revealed several important outcomes regarding gender bias in both humans and models:
Human Bias vs. Model Bias: Humans showed a greater tendency for gender bias with natural sentences compared to synthetic ones, indicating that the nature of the content can significantly influence bias. For synthetic sentences, models displayed stronger biases.
Influence of Time Constraints: As participants were given less time to read sentences, their gender bias increased. This finding highlights how limited processing time can exacerbate biases in decision-making.
Response Time Trends: The amount of time taken by participants to make decisions was related to the presence of bias. Longer response times were observed when distinguishing pronouns from distractors, suggesting that quick judgments can lead to biased decisions.
Comparison of Errors: When looking at errors made by both humans and models, it was noted that models tended to err more with professions strongly associated with a specific gender. In contrast, humans made errors across a wider range of professions.
Differences in Performance: Overall, models showed less accuracy on real-world sentences, while humans tended to perform better, indicating that humans may draw on common sense reasoning more effectively in natural contexts.
Conclusion
The research contributes to the understanding of how biases operate within both human decision-making and language models. The parallels found between the two suggest that biases are not only inherent in language models but reflect broader societal issues present in human cognition.
While language models are trained on large datasets that may contain biases, the individuals interpreting language bring their own biases into play. By understanding these similarities, steps can be taken to reduce bias in language models, which could lead to fairer and more accurate language processing systems.
Future research could expand on these findings by examining different languages, exploring more diverse datasets, and looking into how biases may differ across cultural contexts. Additionally, evaluating how to incorporate features that mitigate these biases in models could enhance fairness in language understanding technologies.
Implications for the Future
The ongoing exploration of gender bias in language processing is vital, especially as models become increasingly integrated into everyday applications like chatbots, translation services, and content generation. Recognizing and addressing biases will not only improve the functionality of these systems but also promote a more equitable digital environment.
Understanding the cognitive processes that lead to bias, both in humans and machines, can lead to better design and training of models. It emphasizes the importance of considering ethical implications when developing artificial intelligence technologies.
As we advance toward more sophisticated systems, maintaining awareness of how biases influence language understanding will be crucial for ensuring that artificial intelligence serves all users fairly and effectively.
Title: Comparing Humans and Models on a Similar Scale: Towards Cognitive Gender Bias Evaluation in Coreference Resolution
Abstract: Spurious correlations were found to be an important factor explaining model performance in various NLP tasks (e.g., gender or racial artifacts), often considered to be ''shortcuts'' to the actual task. However, humans tend to similarly make quick (and sometimes wrong) predictions based on societal and cognitive presuppositions. In this work we address the question: can we quantify the extent to which model biases reflect human behaviour? Answering this question will help shed light on model performance and provide meaningful comparisons against humans. We approach this question through the lens of the dual-process theory for human decision-making. This theory differentiates between an automatic unconscious (and sometimes biased) ''fast system'' and a ''slow system'', which when triggered may revisit earlier automatic reactions. We make several observations from two crowdsourcing experiments of gender bias in coreference resolution, using self-paced reading to study the ''fast'' system, and question answering to study the ''slow'' system under a constrained time setting. On real-world data humans make $\sim$3\% more gender-biased decisions compared to models, while on synthetic data models are $\sim$12\% more biased.
Authors: Gili Lior, Gabriel Stanovsky
Last Update: 2023-05-24 00:00:00
Language: English
Source URL: https://arxiv.org/abs/2305.15389
Source PDF: https://arxiv.org/pdf/2305.15389
Licence: https://creativecommons.org/licenses/by-nc-sa/4.0/
Changes: This summary was created with assistance from AI and may have inaccuracies. For accurate information, please refer to the original source documents linked here.
Thank you to arxiv for use of its open access interoperability.
Reference Links
- https://www.unf.edu/lgbtqcenter/Pronouns.aspx
- https://www.bls.gov/cps/cpsaat11.htm
- https://leaderboard.allenai.org/
- https://github.com/addrummond/ibex
- https://about.citiprogram.org/course/human-subjects-research-social-behavioral-educational-sbe-refresher-1/
- https://github.com/julianmichael/qasrl-modeling
- https://arxiv.org/pdf/2111.07997.pdf
- https://aclanthology.org/2021.eacl-main.137.pdf
- https://brown.edu/Research/AI/files/pubs/wsdm18.pdf
- https://aclanthology.org/2022.cmcl-1.9/
- https://github.com/SLAB-NLP/Cog-GB-Eval
- https://github.com/yuvalkirstain/s2e-coref
- https://anvil.works/
- https://link.springer.com/chapter/10.1007/978-981-10-7563-6_53
- https://www.latex-project.org/help/documentation/encguide.pdf