Skip to content

An AI bias detection and mitigation tool for ML engineers and organizations, the result of the Entourage Penguins 48-hour team effort at Hackathon for Good 6

Notifications You must be signed in to change notification settings

sabinagio/AI-bias-detection-and-mitigation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

PULSE - AI bias detection and mitigation in sentiment analysis

This project was created in 48 hours by the Entourage Penguins team (Camille Coeurjoly, Leo Stahlschmidt, and Sabina Firtala) at the 6th edition of the Hackathon for Good.

Problem Statement

Currently, AI developers face considerable challenges in identifying and rectifying biases in AI datasets. Traditional methods for bias detection and correction are often time-consuming and complex, making them impractical for rapid development cycles. This delay in addressing biases can lead to their perpetuation in AI outputs, reinforcing existing prejudices and inequalities. The challenge is to streamline this process, enabling swift and efficient bias mitigation.

Read more about the challenge here.

Solution

Our team created PULSE, a Streamlit application which detects and mitigates racist AI modelling results for social media content moderation. PULSE detects bias in model outputs statistically and by calculating the bias AUC scores as devised by Google Jigsaw and it mitigates it using both weight correction and adversarial debiasing. To detect whether a text includes a racial subgroup, we used Holistic Bias descriptors developed by Meta. We present the CrowdFlower dataset as our use-case and welcome any suggestions for improvement.

Note: The PULSE version currently in GitHub may still be missing some functionality. Please contact the repository owner (Sabina Firtala) regarding issues.

Data

@inproceedings{hateoffensive, title = {Automated Hate Speech Detection and the Problem of Offensive Language}, author = {Davidson, Thomas and Warmsley, Dana and Macy, Michael and Weber, Ingmar}, booktitle = {Proceedings of the 11th International AAAI Conference on Web and Social Media}, series = {ICWSM '17}, year = {2017}, location = {Montreal, Canada}, pages = {512-515} }

About

An AI bias detection and mitigation tool for ML engineers and organizations, the result of the Entourage Penguins 48-hour team effort at Hackathon for Good 6

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published