Algorithmic fairness deals with building “fair” 🤖 Machine Learning models. The main definitions of fairness group fairness and individual fairness; the former considers statistics across subgroups in the populations, and the latter considers each person individually.
However, among both definitions, choices for what constitutes “harm”—and thus how to equalize harm and guarantee fairness—is subjective and domain-specific, as seen in the 🧭 COMPAS battle. Even worse, by the 🪶 Multi-Fairness Unsatisfiability Theorem, it’s often impossible to satisfy multiple definitions of fairness at once.
Machine Learning Bias
Standard machine learning optimization doesn’t ensure fairness. Simply minimizing error often gives us models that are more effective on some groups than others. Note that for this reason, sometimes a model trained on datasets without sensitive attributes (like race) can be unfair to these groups—on the contrary, a model that considers such attributes may ironically be more fair.
There are many ways standard optimization practices may go wrong.
- Less data on certain subgroups.
- Different groups have different distributions.
- Features are less predictive on some groups.
- Some groups are inherently less predictable (for example, face recognition with or without facial hair).
- Data is biased.
Group Fairness
Group fairness algorithms aim to reduce the differences in harm across different groups. Two common methods are 🔩 Bolt-On Bias Mitigation and the 🔮 Oracle Fairness Approach.
Individual Fairness
Individual fairness is a bit harder to define. We can’t treat each individual as their own group since their error would be
Metric Fairness
Metric fairness applies the idea that “similar individuals should get treated similarly.” Specifically, it defines some distance metric
Subgroup Fairness
Subgroup fairness notes that if we achieve group fairness, subgroups defined by membership in intersecting groups (“disabled Hispanic women over age 55,” for example) can still face discrimination. To solve this, we can employ a similar learner-regulator framework from the 🔮 Oracle Fairness Approach to enforce violations on groups
Average Individual Fairness
Average individual fairness considers cases where our model
Fairness Elicitation
Fairness elicitation goes beyond simple definitions of fairness and instead elicits empirical judgements from other people. These judgement have the form ”
If we have the outcome data
Minimax Group Fairness
Minimax group fairness notes that for some methods that aim to equalize harm, we may be needlessly inflating harm for the advantaged subgroups to ensure equality. An alternative guarantee is to promise that the error on each group is beneath a threshold—but no promise that errors are equal across groups. This gives us the minimax problem
Bias Bounties
💰 Bias Bounties is an approach that invites the public to help improve the model. Similar to bug bounties, we reward people for finding biases and fixing them.