15 Fairness
15 Fairness
Justin Johnson & David Fouhey March 25, 2021EECS 442 WI 2021
Harms of ML Systems
• Allocative Harms : a ML system that is used to allocate resource can
allocate it unfairly or perpetuate inequality
• Sentencing criminals, loan applications, insurance rates, job applications
Source: https://www.reddit.com/r/europe/comments/m9uphb/hungarian_has_no_gendered_pronouns_so_google
Representational Harms
Justin Johnson & David Fouhey March 25, 2021EECS 442 WI 2021
Representational Harms
Justin Johnson & David Fouhey March 25, 2021EECS 442 WI 2021
Representational Harms
Task: Gender Classification
Input: RGB Image
Output: {Man, Woman} Prediction
Buolamwini and Gebru, “Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification”, FAT* 2018
Fairness
• Intuition: are people are being treated equally?
Justin Johnson & David Fouhey March 25, 2021EECS 442 WI 2021
Case Study: COMPAS
• 2016 ProPublica article analyzed COMPAS scores for >7000 people
arrested in Broward county, Florida
• Question: How many of these people ended up committing new
crimes within 2 years?
Justin Johnson & David Fouhey March 25, 2021EECS 442 WI 2021
Error Metrics: Different Stakeholders
Justin Johnson & David Fouhey March 25, 2021EECS 442 WI 2021
Case Study: COMPAS
Justin Johnson & David Fouhey March 25, 2021EECS 442 WI 2021
Case Study: COMPAS
• COMPAS gives very different outcomes for white vs black defendants,
but it does not use race as an input to the algorithm!
https://developers.google.com/machine-learning/crash-course/fairness/evaluating-for-bias
Demographic Parity
• Idea: compare the admissions rate for the majority group and the
minority group
• If it is (approximately) equal, a student's chance of being admitted to
the university doesn't vary by demographic group.
https://developers.google.com/machine-learning/crash-course/fairness/demographic-parity
Demographic Parity
https://developers.google.com/machine-learning/crash-course/fairness/demographic-parity
Demographic Parity Pros and Cons
• Demographic parity is independent from the target variable, there is
no need to have access to its data to measure and correct bias. So it is
suitable for applications when the target is hard to qualify
(employment qualification, credit default, justice, etc.)
https://developers.google.com/machine-learning/crash-course/fairness/demographic-parity
Demographic Parity Issues
https://developers.google.com/machine-learning/crash-course/fairness/demographic-parity
Demographic Parity Issues
https://developers.google.com/machine-learning/crash-course/fairness/demographic-parity
Equality of Opportunity
• Idea: compare the acceptance rates for just the qualified candidates
in the majority group and the minority group. If the acceptance rates
for qualified students in both groups are equal, the model exhibits
equality of opportunity
https://developers.google.com/machine-learning/crash-course/fairness/types-of-bias
Bias Taxonomy
• Historical bias occurs when historical data reflects inequities that
existed in the world at that time.
• A city housing dataset from the 1960s contains home-price data that
reflects discriminatory lending practices in effect during that decade.
https://developers.google.com/machine-learning/crash-course/fairness/types-of-bias
Bias Taxonomy
• Coverage bias occurs if data is not selected in a representative
fashion.
https://developers.google.com/machine-learning/crash-course/fairness/types-of-bias
Bias Taxonomy
• Non-response bias (also known as participation bias) occurs if data
ends up being unrepresentative due to participation gaps in the data-
collection process.
https://developers.google.com/machine-learning/crash-course/fairness/types-of-bias
Bias Taxonomy
• Sampling bias occurs if proper randomization is not used during data
collection.
https://developers.google.com/machine-learning/crash-course/fairness/types-of-bias
Bias Taxonomy
• In-group bias is a preference for members of your own group you also
belong, or for characteristics that you also share.
https://developers.google.com/machine-learning/crash-course/fairness/types-of-bias
Bias Taxonomy
• Out-group homogeneity bias is a tendency to stereotype individual
members of a group to which you do not belong, or to see their
characteristics as more uniform.
https://developers.google.com/machine-learning/crash-course/fairness/types-of-bias
Bias Taxonomy
• Implicit bias occurs when assumptions are made based on one's own
model of thinking and personal experiences that don't necessarily
apply more generally.
https://developers.google.com/machine-learning/crash-course/fairness/types-of-bias
Bias Taxonomy
• Confirmation bias occurs when model builders unconsciously process
data in ways that affirm pre-existing beliefs and hypotheses.
https://developers.google.com/machine-learning/crash-course/fairness/types-of-bias
Bias Taxonomy
• Experimenter's bias occurs when a model builder keeps training a model
until it produces a result that aligns with their original hypothesis.
https://developers.google.com/machine-learning/crash-course/fairness/types-of-bias
Interlude
• Cybersecurity and VirusTotal
Mitigating Bias
• In many cases training dataset is biased, and collecting a less biased
dataset is not always possible.
• Post processing
• Adjust the outputs of the model to abide by fairness criteria.
Post Processing Techniques for Bias Mitigation
• A standard classifier is first learned, and then the learned classifier is
modified to satisfy a fairness constraint
• Oblivious assumption: fair class labels are determined based only on
labels of a standard classifier and a sensitive value
Fairness Tree