Title
Monolingual and Multilingual Reduction of Gender Bias in Contextualized Representations.
Abstract
Pretrained language models (PLMs) learn stereotypes held by humans and reflected in text from their training corpora, including gender bias. When PLMs are used for downstream tasks such as picking candidates for a job, people’s lives can be negatively affected by these learned stereotypes. Prior work usually identifies a linear gender subspace and removes gender information by eliminating the subspace. Following this line of work, we propose to use DensRay, an analytical method for obtaining interpretable dense subspaces. We show that DensRay performs on-par with prior approaches, but provide arguments that it is more robust and provide indications that it preserves language model performance better. By applying DensRay to attention heads and layers of BERT we show that gender information is spread across all attention heads and most of the layers. Also we show that DensRay can obtain gender bias scores on both token and sentence levels. Finally, we demonstrate that we can remove bias multilingually, e.g., from Chinese, using only English training data.
Year
DOI
Venue
2020
10.5282/UBM/EPUB.74040
COLING
DocType
Volume
Citations 
Conference
2020.coling-main
0
PageRank 
References 
Authors
0.34
0
3
Name
Order
Citations
PageRank
Sheng Liang16613.09
Philipp Dufter214.74
Hinrich Schütze32113362.21