Module1 Lecture 1
Module1 Lecture 1
Hima Lakkaraju
Assistant Professor
Harvard Business School + Computer Science
Background
2
Motivation
[ Weller 2017 ]
Motivation: Why Model Understanding?
This prediction is
Defendant Details biased. Race and
gender are being
used to make the
prediction!!
Model Understanding
Model understanding facilitates bias detection.
Race
Crimes
Gender
Predictive
Prediction = Risky to Release
Model
Increase
Loan Applicant Details
salary by I have some means
50K + pay for recourse. Let me
credit card go and work on my
promotion and pay
bills on time
my bills on time.
Model understanding helps
for next 3 provide recourse to individuals
months
who are adversely to
affected by model predictions.
get a loan
Predictive
Prediction = Denied Loan
Model
Loan Applicant
Motivation: Why Model Understanding?
Model Understanding
Patient Data This model is using
If gender = female, irrelevant features when
if ID_num > 200, then sick predicting on female
subpopulation. I should
If gender = male, not trust its predictions
25, Female, Cold if cold = true and cough = true, then for that group.
32, Male, No
31, Male, Cough
Model understanding helps assess
sick if and when to trust
model predictions when making decisions.
Predictions
Healthy
Sick
Predictive Sick
Model .
.
Healthy
Healthy 7
Sick
Motivation: Why Model Understanding?
Predictions
Healthy
Sick
Predictive Sick
Model .
.
Healthy
Healthy 8
Sick
Motivation: Why Model Understanding?
Utility Stakeholders
Explainer
Example
[ Cireşan et. al. 2012, Caruana et. al. 2006, Frosst et. al. 2017, Stewart 2020]
Inherently Interpretable Models vs.
Post hoc Explanations
18
Motivation for Interpretability
19
Prior Work: Defining and Measuring
Interpretability
Little consensus on what interpretability is and how
to evaluate it
20
Prior Work: Defining and Measuring
Interpretability
Evaluate in the context of an application
If a system is useful in a practical application or a
simplified version, it must be interpretable
21
Lack of Rigor?
Yes and No
Previous notions are reasonable
Important to formalize these notions!!!
However,
23
When and Why Interpretability?
24
When and Why Interpretability?
Incompleteness ≠ Uncertainty
Uncertainty can be quantified
E.g., trying to learn from a small dataset (uncertainty)
25
Incompleteness: Illustrative Examples
Scientific Knowledge
E.g., understanding the characteristics of a large dataset
Goal is abstract
Safety
End to end system is never completely testable
Not possible to check all possible inputs
Ethics
Guard against certain kinds of discrimination which are too
abstract to be encoded
No idea about the nature of discrimination beforehand
26
Taxonomy of Interpretability Evaluation
27
Application-grounded evaluation
Potential experiments
Pairwise comparisons
Simulate the model output
What changes should be made to input to change the
output?
29
Functionally-grounded evaluation
Potential experiments
Complexity (of a decision tree) compared to other other
models of the same (similar) class
How many levels? How many rules?
30
Open Problems: Design Issues
31
Taxonomy based on applications/tasks
Degree of Incompleteness
What part of the problem is incomplete? How incomplete
is it?
Incomplete inputs or constraints or costs?
Time Constraints
How much time can the user spend to understand
explanation?
32
Taxonomy based on applications/tasks
33
Taxonomy based on methods
34
Taxonomy based on methods
Level of compositionality:
Are the basic units organized in a structured way?
How do the basic units compose to form higher order
units?
Uncertainty:
What kind of uncertainty is captured by the methods?
How easy is it for humans to process uncertainty?
35
Questions??
Relevant Conferences to Explore
ICML
NeurIPS
ICLR
UAI
AISTATS
KDD
AAAI
FAccT
AIES
CHI
CSCW
HCOMP 37
Breakout Groups