AUB ScholarWorks

Towards a Robust Gender Bias Evaluation in NLP

Show simple item record

dc.contributor.advisor Khreich, Wael
dc.contributor.author Barza, Kenny
dc.date.accessioned 2023-02-10T11:41:08Z
dc.date.available 2023-02-10T11:41:08Z
dc.date.issued 2/10/2023
dc.date.submitted 2/6/2023
dc.identifier.uri http://hdl.handle.net/10938/23968
dc.description.abstract With the advent of deep learning technology, Natural Language Processing (NLP) has made remarkable progress. Deep learning models have improved the performance of many NLP tasks such as text summarization, translation, and sentiment analysis. However, NLP models have been shown to present gender biases, which can be detrimental to decision-making. As a result, assessing the gender bias of these models before deploying them is a must. We develop the Gender Bias Evaluation Framework (GBEF), a framework that measures gender bias in Masked Language Models (MLMs). The GBEF consists of two approaches. Each approach uses preconstructed data and a gender bias metric. The first evaluation approach is called the Sentence-Based Evaluation (SBE) and it can assess gender bias in three different categories: occupation bias, benevolent sexism, and hostile sexism. The second approach is called the Template-Based Evaluation (TBE) and will be used for a more accurate assessment of the counterfactual data substitution debiasing technique, a technique that relies on balancing female-related words and male-related words in the training corpus. We first use the SBE to quantify gender bias in different BERT models and show that BERTlarge is the most biased model while RoBERTalarge is the least gender biased one. The SBE was also used to quantify gender bias in corpora. We develop a new method for this task that relies on fine-tuning BERT for the masked language model task on the corpus on which we want to measure the bias. We compare Jigsaw’s toxic comments with Jigsaw’s severe toxic comments and reveal that the latter presents a higher degree of gender bias. Finally, the TBE was able to shed light on the issues of the debiasing technique that relies on fine-tuning BERT on a counterfactual data substituted corpus. While this technique was able to reduce gender bias in BERT at a high level, we show with the TBE that the model is simply treating male and female-related pronouns as equal, which may be problematic when it comes to gender-related words (e.g., pregnant). We propose a solution to this problem by including sentences with gender-related words in the training corpus. The inclusion of these sentences in the training corpus allowed the debiased version of BERT to associate gender-related words with the right gender. We believe that our proposed evaluation framework will aid in a more accurate assessment of the gender bias in different MLMs improving fairness in artificial intelligence.
dc.language.iso en
dc.subject Natural Language Processing
dc.subject Gender Bias
dc.subject Ethical Artificial Intelligence
dc.subject Masked Language Model
dc.subject BERT
dc.title Towards a Robust Gender Bias Evaluation in NLP
dc.type Thesis
dc.contributor.department Business Information and Decision Systems
dc.contributor.faculty Suliman S. Olayan School of Business
dc.contributor.institution American University of Beirut
dc.contributor.commembers Mouassi, Lama
dc.contributor.commembers Zablith, Fouad
dc.contributor.commembers Taleb, Sirine
dc.contributor.degree MS
dc.contributor.AUBidnumber 202123106


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search AUB ScholarWorks


Browse

My Account