Tackling Hate Speech in Devanagari Languages
A study on using AI to detect hate speech in Hindi and Nepali.
― 5 min read
Table of Contents
In today’s digital world, the spread of Hate Speech online is a serious issue. It can lead to real-world harm, especially to vulnerable communities. While this challenge affects many places, it's particularly noticeable in languages that use the Devanagari script, such as Hindi and Nepali. There aren't many tools or resources available to tackle hate speech in these languages, making the problem more difficult to address.
The Importance of Detecting Hate Speech
Hate speech can cause a lot of damage, which is why detecting it is crucial. The online world is like a big party where some people always try to ruin the fun for others. When hate speech is detected early, it can help reduce its spread and impact. Unfortunately, detecting hate speech in languages like Hindi and Nepali is tough.
Large Language Models?
What areLarge Language Models (LLMs) are like super-smart robots that can understand and use human language. They are built on lots of data and can perform various language tasks. However, they usually need a lot of resources to be tuned properly, which can be tough to manage in low-resource languages. Imagine trying to get a giant elephant to dance; it’s not an easy task!
The Challenge with Traditional Techniques
Traditional methods for training these models can be expensive. It’s like trying to buy shoes for a giant-you need a lot of materials and a big budget! This can be especially hard for languages that don’t have as many resources available. So, researchers are on the lookout for smarter ways to fine-tune these models without breaking the bank.
Parameter Efficient Fine-Tuning (PEFT)
This is where Parameter Efficient Fine-Tuning (PEFT) comes into play. Rather than tuning the whole elephant, we just make small adjustments that keep it dancing gracefully. PEFT allows us to fine-tune just a part of the model’s parameters, making it more suitable for languages with fewer resources.
LoRA: A Smart Approach
One technique under PEFT is called LoRA (Low-Rank Adaptation). Imagine LoRA as a tiny mechanic working on a big machine. It focuses on adjusting just a few areas, which not only lowers the cost but also helps the machine run smoothly without delays. This saves time and resources while maintaining efficiency.
The Study: Detecting Hate Speech in Devanagari Languages
This study focuses on detecting hate speech in Hindi and Nepali using LLMs. The researchers set up a system to analyze the text in these languages. It's like having a friendly robot that can spot troublemakers at a party before they start causing chaos.
Datasets
TheTo train the LLMs, they used a dataset containing thousands of examples of text. This text was taken from various sources, including social media posts and news articles. Unfortunately, they found that most of the texts were non-hate speech, creating an imbalance. This is like having a jar full of jellybeans, where 90% are red and only 10% are green. It makes it hard for the robot to learn which ones are bad!
Training the Models
The study involved testing various LLMs on this dataset. Specifically, they looked at how well different models performed in detecting hate speech and identifying its targets. This means not only figuring out if a piece of text contained hate speech but also if it was aimed at a person, organization, or community.
Results and Analysis
After running the tests, the researchers found that one model, called Nemo, performed the best in both tasks. It's like finding out that the little engine that could was actually a race car! Despite having fewer parameters than some other models, Nemo managed to deliver outstanding results.
Class Imbalance Issues
A key part of their findings was that the model worked significantly better at identifying non-hate speech than hate speech. This was largely due to the imbalance in the training data. The more hate speech they fed it, the better it would get at recognizing it, but they had a much larger number of non-hate speech examples. So it’s like trying to teach a dog to bark when it’s surrounded by a bunch of silent cats!
Target Identification Challenges
When it came to identifying targets of hate speech, the researchers noticed another issue. The model struggled to recognize hate speech directed at communities. This highlights the challenges of classifying targets when some categories have fewer examples.
Conclusion and Future Work
In conclusion, the study showed that using LLMs with efficient fine-tuning methods can help detect hate speech in languages that often get overlooked. While they achieved good performance, there are still challenges ahead, particularly with imbalanced datasets. Moving forward, researchers plan to develop techniques to create more balanced datasets, which would help improve the model's accuracy.
Ethical Considerations
Detecting hate speech isn’t just a technical issue; it’s also an ethical one. The researchers noted that the models can have biases, so it’s essential to have human reviews before making any decisions based on the models’ predictions. This ensures that we don’t accidentally accuse an innocent jellybean of being a troublemaker.
The Bigger Picture
As we move further into a digital age, developing tools to detect hate speech is necessary to create a safer online environment. The hope is that with continued research and better resources, we can tackle these problems more effectively, helping to keep the online party enjoyable for everyone. So, let’s continue to build those smart robots and give them the tools they need to keep the peace!
Title: LLMsAgainstHate @ NLU of Devanagari Script Languages 2025: Hate Speech Detection and Target Identification in Devanagari Languages via Parameter Efficient Fine-Tuning of LLMs
Abstract: The detection of hate speech has become increasingly important in combating online hostility and its real-world consequences. Despite recent advancements, there is limited research addressing hate speech detection in Devanagari-scripted languages, where resources and tools are scarce. While large language models (LLMs) have shown promise in language-related tasks, traditional fine-tuning approaches are often infeasible given the size of the models. In this paper, we propose a Parameter Efficient Fine tuning (PEFT) based solution for hate speech detection and target identification. We evaluate multiple LLMs on the Devanagari dataset provided by (Thapa et al., 2025), which contains annotated instances in 2 languages - Hindi and Nepali. The results demonstrate the efficacy of our approach in handling Devanagari-scripted content.
Authors: Rushendra Sidibomma, Pransh Patwa, Parth Patwa, Aman Chadha, Vinija Jain, Amitava Das
Last Update: 2024-12-26 00:00:00
Language: English
Source URL: https://arxiv.org/abs/2412.17131
Source PDF: https://arxiv.org/pdf/2412.17131
Licence: https://creativecommons.org/licenses/by/4.0/
Changes: This summary was created with assistance from AI and may have inaccuracies. For accurate information, please refer to the original source documents linked here.
Thank you to arxiv for use of its open access interoperability.