Bias in Language Models: The Math Anxiety Issue
Examining how language models reflect societal biases in math and STEM.
― 6 min read
Table of Contents
Large language models (LLMs) like GPT-3, ChatGPT, and GPT-4 have become a big part of our daily lives. We use them for many tasks, from writing emails to answering questions. However, it is crucial to look at the biases in the outputs of these models to prevent the spread of harmful stereotypes. Bias can often reflect how society thinks about certain topics. One area of concern is the widespread anxiety people feel about math and STEM subjects. This article looks into how different LLMs perceive math and STEM fields and how these perceptions might mirror the anxieties of high school students.
The Rise of Language Models
LLMs have gained a lot of attention for their ability to perform a variety of language tasks. They are trained on vast amounts of text data, allowing them to recognize patterns in language. This makes them seem nearly human in their ability to generate text and solve problems. As these models become more integrated into our lives, the need to understand their behavior and the risks involved becomes even more pressing. Research in this area has increased, with debates about whether these models actually understand language or merely simulate it.
Understanding Bias in Language Models
Bias is a significant issue when it comes to LLMs. It refers to the misrepresentation of reality, favoring certain groups or ideas while perpetuating stereotypes. LLMs learn from the text data they are trained on, which often contains biases present in society. This means that the outputs generated by these models can reflect and even amplify existing stereotypes.
As LLMs become more common, it is essential to look closely at the biases they produce. These models act as mirrors, reflecting not only our individual biases but also societal views and trends. One area where these biases manifest is in affective bias, which is the tendency to focus more on negative events than positive ones. For example, people may attribute negative attitudes to math, leading to a cycle of negative perceptions.
Math Anxiety and Its Impact
In many societies, math is often viewed as a challenging subject, causing anxiety for students. This anxiety can begin in childhood and continue into adulthood. Many people identify as "not math people," which can lead to avoidance of math-related tasks. This avoidance can have significant effects on academic and professional paths, particularly in STEM fields.
The feelings around math can be transmitted from teachers and parents to children, much like how LLMs absorb biases from the data they are trained on. This means that the negative perceptions surrounding math can influence how students see the subject and themselves. Understanding these biases in LLMs is important to ensure that positive attitudes towards math can be encouraged, particularly since negative views can lead to a gender gap in STEM careers.
The Role of Behavioral Forma Mentis Networks
To study how LLMs perceive math and STEM, researchers can use a method called behavioral forma mentis networks (BFMNs). This approach helps show how certain concepts are associated in people's minds. By building a network of these associations, researchers can better grasp how LLMs frame math and STEM disciplines.
In this study, the authors looked at Responses from LLMs to specific cue words related to math and STEM fields. They compared the findings with responses from high school students to see how perceptions aligned or differed. This approach helps reveal the attitudes reflected in the outputs of these models.
Methods Used for Investigation
In the study, various cue words were provided to the LLMs, and they were asked to generate three words that came to mind along with a rating for each word on a scale of 1 to 5, indicating how positive or negative each word felt. For comparison, high school students were also asked to provide similar responses. This process resulted in datasets that could be analyzed to understand how both students and LLMs frame key concepts like math and science.
Understanding the Results
When analyzing the responses from the different LLMs, researchers found that GPT-3 and ChatGPT had a negative perception of math. Both models produced fewer positive associations and framed math using words that had negative connotations. For example, terms like "difficult" and "boring" were common. In contrast, the more advanced GPT-4 revealed a shift in this perception, producing a more neutral or even positive take on math, suggesting an evolution in how these models understand the subject.
The results from these interactions show that LLMs are not neutral in their framing of academic subjects. They tend to reflect the negative attitudes found in high school students about math and STEM. This bias is concerning, especially as more students and individuals interact with these models.
Comparing LLMs with Human Responses
The study also compared the responses of LLMs with those of high school students. Interestingly, it was found that both LLMs and students viewed math negatively, but as the models advanced, like with GPT-4, there was a noticeable change towards more neutral or positive associations. For instance, GPT-4 linked math with various positive concepts, while earlier models like GPT-3 and ChatGPT did not.
These findings suggest that the latest language models might have the potential to reduce the negative framing that has been characteristic of earlier versions. This is crucial because negative associations can influence students' feelings about math and lead to avoidance, impacting their educational and career choices.
Emotional Analysis and Visualization
To further understand how math and STEM concepts are perceived, researchers employed a circumplex model of affect. This model plots emotions on a graph, helping visualize how different words and concepts relate to feelings of positivity or negativity and how they energize or calm individuals. The results indicated that while science often inspired positive feelings, math retained a largely negative framing across the models, particularly in earlier versions.
By looking at the emotional context of the words associated with math, researchers could better grasp why students feel anxious about the subject. The emotional responses to key concepts reveal deep-rooted societal attitudes, which these models mirror in their outputs.
Conclusion
In summary, the study of LLMs like GPT-3, ChatGPT, and GPT-4 shows that they often reflect society’s biases, particularly regarding math and STEM subjects. Math anxiety is a widespread issue, echoed in the outputs of these models, which can perpetuate negative perceptions. However, advancements in LLMs suggest a shift towards less biased views, particularly in more recent models like GPT-4.
As these models continue to be used in educational settings, it’s essential to be aware of the biases they may promote. Encouraging positive perceptions of math and STEM can help combat math anxiety and inspire more students to pursue these fields. Additionally, further research into the influences of language models will be necessary to ensure they contribute positively to learning and development without reinforcing negative stereotypes.
Recognizing the potential effects of LLMs on attitudes toward math is essential for educators, parents, and policymakers as they work to create environments that foster interest and confidence in these important subjects.
Title: Cognitive network science reveals bias in GPT-3, ChatGPT, and GPT-4 mirroring math anxiety in high-school students
Abstract: Large language models are becoming increasingly integrated into our lives. Hence, it is important to understand the biases present in their outputs in order to avoid perpetuating harmful stereotypes, which originate in our own flawed ways of thinking. This challenge requires developing new benchmarks and methods for quantifying affective and semantic bias, keeping in mind that LLMs act as psycho-social mirrors that reflect the views and tendencies that are prevalent in society. One such tendency that has harmful negative effects is the global phenomenon of anxiety toward math and STEM subjects. Here, we investigate perceptions of math and STEM fields provided by cutting-edge language models, namely GPT-3, Chat-GPT, and GPT-4, by applying an approach from network science and cognitive psychology. Specifically, we use behavioral forma mentis networks (BFMNs) to understand how these LLMs frame math and STEM disciplines in relation to other concepts. We use data obtained by probing the three LLMs in a language generation task that has previously been applied to humans. Our findings indicate that LLMs have an overall negative perception of math and STEM fields, with math being perceived most negatively. We observe significant differences across the three LLMs. We observe that newer versions (i.e. GPT-4) produce richer, more complex perceptions as well as less negative perceptions compared to older versions and N=159 high-school students. These findings suggest that advances in the architecture of LLMs may lead to increasingly less biased models that could even perhaps someday aid in reducing harmful stereotypes in society rather than perpetuating them.
Authors: Katherine Abramski, Salvatore Citraro, Luigi Lombardi, Giulio Rossetti, Massimo Stella
Last Update: 2023-05-22 00:00:00
Language: English
Source URL: https://arxiv.org/abs/2305.18320
Source PDF: https://arxiv.org/pdf/2305.18320
Licence: https://creativecommons.org/licenses/by/4.0/
Changes: This summary was created with assistance from AI and may have inaccuracies. For accurate information, please refer to the original source documents linked here.
Thank you to arxiv for use of its open access interoperability.