Leveraging Bias in Pre-trained Word Embeddings for Unsupervised Microaggression Detection

Microaggressions are subtle manifestations of bias (Breitfeller et al. 2019). These demonstrations of bias can often be classified as a subset of abusive language. However, not much focus has been placed on the recognition of these instances. As a result, limited data is available on the topic, and...

Full description

Saved in:
Bibliographic Details
Main Authors: Tolúlọpẹ́ Ògúnrẹ̀mí, Valerio Basile, Tommaso Caselli
Format: Article
Language:English
Published: Accademia University Press 2022-12-01
Series:IJCoL
Online Access:https://journals.openedition.org/ijcol/1066
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Microaggressions are subtle manifestations of bias (Breitfeller et al. 2019). These demonstrations of bias can often be classified as a subset of abusive language. However, not much focus has been placed on the recognition of these instances. As a result, limited data is available on the topic, and only in English. Being able to detect microaggressions without the need for labeled data would be advantageous since it would allow content moderation also for languages lacking annotated data. In this study, we introduce an unsupervised method to detect microaggressions in natural language expressions. The algorithm relies on pre-trained word-embeddings, leveraging the bias encoded in the model in order to detect microaggressions in unseen textual instances. We test the method on a dataset of racial and gender-based microaggressions, reporting promising results. We further run the algorithm on out-of-domain unseen data with the purpose of bootstrapping corpora of microaggressions “in the wild”, perform a pilot experiment with prompt-based learning, and discuss the benefits and drawbacks of our proposed method.1
ISSN:2499-4553