Hate Speech Detection Using Large Language Models: A Comprehensive Review

The widespread use of social media and other online platforms has facilitated unprecedented communication and information exchange. However, it has also led to the spread of hate speech and poses serious challenges to societal harmony as well as individual well-being. Traditional methods for detecti...

Full description

Saved in:
Bibliographic Details
Main Authors: Aish Albladi, Minarul Islam, Amit Das, Maryam Bigonah, Zheng Zhang, Fatemeh Jamshidi, Mostafa Rahgouy, Nilanjana Raychawdhary, Daniela Marghitu, Cheryl Seals
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10848067/
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:The widespread use of social media and other online platforms has facilitated unprecedented communication and information exchange. However, it has also led to the spread of hate speech and poses serious challenges to societal harmony as well as individual well-being. Traditional methods for detecting hate speech, such as keyword matching, rule-based systems, and machine learning algorithms, often struggle to capture the subtle and context-dependent nature of hateful content. This paper provides a comprehensive review of the application of large language models (LLMs) like GPT-3, BERT, and their successors in hate speech detection. We analyze the evolution of LLMs in natural language processing and examine their strengths and limitations in identifying hate speech. Additionally, we address the significant challenges and explore how LLMs method can affect the accuracy and fairness of hate speech detection systems. By synthesizing recent research, this review aims to offer a holistic understanding of the current state-of-the-art methods in hate speech detection utilizing LLMs and to suggest directions for future research that could enhance the efficacy and equity of these systems.
ISSN:2169-3536