Does ChatGPT show gender bias in behavior detection?

Abstract Recent advancements in Large Language Models (LLMs) suggest imminent commercial applications of such AI systems which serve as gateways to interact with technology and the accumulated body of human knowledge. However, whether and how ChatGPT contains bias in behavior detection and thus gene...

Full description

Saved in:
Bibliographic Details
Main Authors: Ji Wu, Yaokang Song, Doris Chenguang Wu
Format: Article
Language:English
Published: Springer Nature 2024-12-01
Series:Humanities & Social Sciences Communications
Online Access:https://doi.org/10.1057/s41599-024-04219-3
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Abstract Recent advancements in Large Language Models (LLMs) suggest imminent commercial applications of such AI systems which serve as gateways to interact with technology and the accumulated body of human knowledge. However, whether and how ChatGPT contains bias in behavior detection and thus generates ethical concerns in decision-making deserves explored. This study therefore aims to evaluate and compare the detection accuracy and gender bias between ChatGPT and such traditional ML methods as Naïve Bayes, SVM, Random Forest, and XGBoost. The experimental results demonstrate that despite of its lower detection accuracy, ChatGPT exhibits less gender bias. Furthermore, we observed that removing gender label features led to an overall reduction in bias for traditional ML methods, while for ChatGPT, the gender bias decreased when labels were provided. These findings shed new light on fairness enhancement and bias mitigation in management decisions when using ChatGPT for behavior detection.
ISSN:2662-9992