Abstract
Christopher Kruegel
Word embeddings such as Word2Vec and GloVe have significantly improved natural language processing tasks by capturing semantic relationships between words. However, they also encode societal biases present in training data, particularly regarding gender, race, and profession. This paper quantifies such biases using established benchmarks like the Word Embedding Association Test (WEAT) and evaluates three mitigation techniques: Hard Debiasing, Gender-Neutral Word Embeddings (GN-GloVe), and Projection Removal. We apply these methods to pre-trained embeddings trained on the Google News and Wikipedia+Gigaword corpora. Bias reduction is measured alongside downstream task performance on analogy completion, sentiment analysis, and named entity recognition (NER). Results show that hard debiasing effectively reduces WEAT scores by over 80%, but sometimes degrades performance on syntactic tasks. GN-GloVe maintains competitive task performance while achieving moderate bias reduction. P
IMPORTANT LINKS
Check Article for
Plagiarism
UPDATES
INDEXED BY: