Skip to content

NauelSerraino/BERT-Gender-Bias

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

24 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

How to quantify gender bias in word embeddings?

This Machine Learning tool helps measure and visualize gender bias in word embeddings.

The tool is based on a pipeline that uses BERT embeddings as a starting point. The pipeline is composed by the following modules:

  1. Logistic Regression - l1 regularization
  2. PCA
  3. Support Vector Classifier

The tool is able provide a visualization instrument to analyze how much word is biased towards male or female gender, here is an example of the visualization:

SVM with output

Performance Metrics:

Metric Value
C (Logistic Regression) 0.175
C (SVM) 0.375
Accuracy 0.7786
Number of Selected Features (pre-PCA) 38

About

Tool to quantify gender bias of words.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages