A comprehensive set of fairness metrics for datasets and machine learning models, explanations for these metrics, and algorithms to mitigate bias in datasets and models.
-
Updated
Nov 13, 2025 - Python
A comprehensive set of fairness metrics for datasets and machine learning models, explanations for these metrics, and algorithms to mitigate bias in datasets and models.
LangFair is a Python library for conducting use-case level LLM bias and fairness assessments
WEFE: The Word Embeddings Fairness Evaluation Framework. WEFE is a framework that standardizes the bias measurement and mitigation in Word Embeddings models. Please feel welcome to open an issue in case you have any questions or a pull request if you want to contribute to the project!
a comprehensive statistical framework for detecting circular reasoning bias in AI algorithm evaluation
Official repository for the paper "ALERT: A Comprehensive Benchmark for Assessing Large Language Models’ Safety through Red Teaming"
Learning to Split for Automatic Bias Detection
Tools for diagnostics and assessment of (machine learning) models
Code & Data for the paper "RedditBias: A Real-World Resource for Bias Evaluation and Debiasing of Conversational Language Models"
This repository contains a console-interface name-ethnicity classifier
"Beyond Skin Tone: A Multidimensional Measure of Apparent Skin Color" (ICCV 2023)
Structured pruning and bias visualization for Large Language Models. Tools for LLM optimization and fairness analysis.
Official code of "Discover and Mitigate Unknown Biases with Debiasing Alternate Networks" (ECCV 2022)
Official code of "Discover the Unknown Biased Attribute of an Image Classifier" (ICCV 2021)
Reveal to Revise: An Explainable AI Life Cycle for Iterative Bias Correction of Deep Models. Paper presented at MICCAI 2023 conference.
Belief-Bias evaluation of local LLMs
CognitiveLens is a Streamlit-powered analytics tool for exploring alignment between human and AI decisions. It visualizes fairness, calibration, and interpretability through metrics like Cohen’s κ, AUC, and Brier score. Designed for ethical AI, bias auditing, and decision transparency in machine learning systems.
"Learning Stable Classifiers by Transferring Unstable Features" ICML 2022
A program to automate testing open source LLMs for their political compass scores
Python library for analyzing data quality and its impact on model performance across classification and object-detection tasks.
Scan your AI/ML models for problems before you put them into production.
Add a description, image, and links to the bias-detection topic page so that developers can more easily learn about it.
To associate your repository with the bias-detection topic, visit your repo's landing page and select "manage topics."