Magnus Nielsen, SODAS, UCPH
This session will be based mainly on the book Fairness and Machine Learning: Limitations and Opportunities (Barocas et al., 2019)
A survey on more aspects and applications can be found in Mehrabi et al. (2021)
Ursula Franklin (Olteanu et al., 2019):
In your opinion, what is fairness in the context of machine learning? What is bias?
Chouldechova (2017):
Nevertheless, we’ll try to turn it into observable statistical measures towards the end
Bias is a broad term with many different interpretations
Some people try to avoid this term due to the ambiguity, including Barocas et al. (2019)
The list is long
Source: Barocas et al. (2019)
Source: Google Translate
Source: Google Translate
Algorithmic fairness is a relatively new concept
Lots of information on fairness, but generally not with respect to algorithms
An ‘absence of any prejudice or favoritism toward an individual or group based on their inherent or acquired characteristics’ (Mehrabi et al., 2021)
We will consider group fairness
For a more thorough walkthrough of moral notions of fairness, see chapter 4 in Barocas et al., 2019
Source: Barocas et al. (2019)
Is it right to deploy a machine learning algorithm in the first place?
Precedes other fairness concerns
Humans are subject to subjectivity, arbitrariness, and inconsistency
Algorithms most commonly replace bureaucracies
Public institutions are in general more regulated than private companies
To increase legitimacy, consider things such as
Automating pre-existing decision-making rules
Learning decision-making rules from data on past decisions in order to automate them
Deriving decision-making rules by learning to predict a target
You’re tasked with creating a new centralized admissions system for higher education using machine learning.
Algorithmic decisions as a process of inductive reasoning
Some issues to consider:
Given a
We can audit the decision-maker!
Ayres & Siegelman (1995)
Bertrand & Mullainathan (2004)
What could be reasons that firms engage in this sort of discrimination?
A wish for null findings implicate no discrimination based on sensitive attribute
Possible to illuminate decision making with partial dependence functions
We have
\[r(x)=P\{Y=1∣X=x\}\]
Based on this risk score, we can calculate a predicted outcome \(Y\)
Can we obtain fairness by withholding information about sensitive attributes?
Source: Barocas et al. (2019)
Fundamentally three types of fairness metrics
These are equalized across groups defined by the sensitive attribute
We will consider the same example for our fairness metrics
Given
Decision to make: Who should get a loan?
Question: What target should we use to create our model?
‘True’ equality
This is also commonly referred to as demographic parity
Easy to work with algorithmically
\[P\{\widehat Y | A = a\} = P\{\widehat Y | A = b\}\]
In practice:
Could independence have adverse consequences for either group in the loans context?
\[P\{\widehat Y | A = a\} \geq P\{\widehat Y | A = b\} - \epsilon\]
\[\frac{P\{\widehat Y | A = a\}}{P\{\widehat Y | A = b\}} \geq 1 - \epsilon\]
The U.S. has an 80% rule to detect discriminatory hiring
No influence if groups differ in covariates and outcomes
Can lead to adverse outcomes, e.g.
Note that minority groups by definition have less training data
Assume that men overall are worse at paying back loans and defaulting is costly
Could this justify some discrimination based on gender?
Equality within error rates
Also known as equalized odds
\[ P \{\widehat Y = 1 | Y = 1, A = a\} = P \{\widehat Y = 1 | Y = 1, A = b\}\]
\[ P \{\widehat Y = 1 | Y = 0, A = a\} = P \{\widehat Y = 1 | Y = 0, A = b\}\]
Equality of
Generally misclassification has a cost, e.g. a lost opportunity
Note that target variables can encode previous inequality and injustice
What are the implications of separation for our loan model?
Source: Barocas et al. (2019)
Tools to achieve red area in figure achievable by a combination of
Equality of either false negative rate or false positive rate
Consider the costs of misclassification and who experiences them
An example: Screening for job interviews
Do you believe a relaxation is appropriate in the loan example?
Given a risk score, groups should not differ in outcomes
If a model predicts a high risk, then it should be the same high risk for both groups
\[P\{Y = 1 ∣ R = r, A = a \} = P\{Y = 1 ∣ R = r, A = b \}\]
A flexible model which uses \(A\) as an input generally (approximately) achieves this
Fairness through awareness
What are the implications of sufficiency for our loan model?
We are generally not able to satisfy all fairness criteria at once!
The proofs for the following three incompatibilities are found in Barocas et al. (2019)
Assume \(A \not\perp Y\)
Then independence and sufficiency cannot both hold
Assume \(Y\) binary, \(A \not\perp Y\) and \(R \not\perp Y\)
Then independence and separation cannot both hold
Assume \(A \not\perp Y\) and all events in the joint distribution \((R, A, Y)\) has positive density
Then separation and sufficiency cannot both hold
When in the process to achieve fairness
Doing only fairness assessment and no mitigation is common
Post-processing is common
Comes with some (near) optimality guarantees
We’ve now looked at different criteria of fairness based on observables
Note that this doesn’t tell us anything about mechanisms or causes
Attempts with causality
Once again remember Chouldechova (2017):
Ayres, I., & Siegelman, P. (1995). Race and gender discrimination in bargaining for a new car. The American Economic Review, 304-321.
Barocas, S., Hardt, M., & Narayanan, A. (2019). Fairness and Machine Learning: Limitations and Opportunities. fairmlbook.org.
Bertrand, M., & Mullainathan, S. (2004). Are Emily and Greg more employable than Lakisha and Jamal? A field experiment on labor market discrimination. American economic review, 94(4), 991-1013.
Chouldechova, A. (2017). Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data, 5(2), 153-163.
Feldman, M., Friedler, S. A., Moeller, J., Scheidegger, C., & Venkatasubramanian, S. (2015, August). Certifying and removing disparate impact. In proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining (pp. 259-268).
Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K., & Galstyan, A. (2021). A survey on bias and fairness in machine learning. ACM Computing Surveys (CSUR), 54(6), 1-35.
Olteanu, A., Castillo, C., Diaz, F., & Kıcıman, E. (2019). Social data: Biases, methodological pitfalls, and ethical boundaries. Frontiers in Big Data, 2, 13.