13 Model Interpretability
13 Model Interpretability
Interpretability
Instructor: Saravanan Thirumuruganathan
Caveat Emptor
• Assumption: You know how to interpret simple ML models
• The ideas in this lecture are quite simple and published > 5 years ago
• So NOT the state of the art which is much much more complex
https://hcixaitutorial.github.io/
Explainable AI
https://hcixaitutorial.github.io/
Why Explainable AI?
• GDPR: Article 22 empowers individuals with the right to demand an explanation of how an
automated system made a decision that affects them.
• Algorithmic Accountability Act 2019: Requires companies to provide an assessment of the risks
posed by the automated decision system to the privacy or security and the risks that
contribute to inaccurate, unfair, biased, or discriminatory decisions impacting consumers
• California Consumer Privacy Act: Requires companies to rethink their approach to capturing,
storing, and sharing personal data to align with the new requirements by January 1, 2020.
• Washington Bill 1655: Establishes guidelines for the use of automated decision systems to
protect consumers, improve transparency, and create more market predictability.
• Massachusetts Bill H.2701: Establishes a commission on automated decision-making,
transparency, fairness, and individual rights.
• Illinois House Bill 3415: States predictive data analytics determining creditworthiness or hiring
decisions may not include information that correlates with the applicant race or zip code.
XAI Tutorial AAAI 2020 Lecue et al
XAI Taxonomy
Guidotti et al. (2018). A survey of methods for explaining black box models. ACM computing surveys (CSUR) .
Post-hoc Global Explanation: Knowledge
Distillation
https://hcixaitutorial.github.io/
Global Surrogate model / Distillation
• Train a complex model M1 on training data D_T
• For dataset D_X, get the prediction from M1
• Choose a simple model M2
• Train M2 using D_X and prediction of M1 on D_X (not ground truth!)
• Make sure the accuracy of M2 on D_X is not too bad compared to M1
Dhurandhar et al. Improving Simple Models with Confidence Profiles. NeurIPS 2018
Glob Exp via Permutation Feature Importance
• Goal: measures the contribution of each feature in a model
Dhurandhar et al. Improving Simple Models with Confidence Profiles. NeurIPS 2018
LIME
Ribeiro, et al. Why should i trust you? Explaining the predictions of any classifier. KDD 2016
Local Approximation via LIME
Gurumoorthy et al. Efficient Data Representation by Selecting Prototypes with Importance Weights”, ICDM 2019
Inspecting Counterfactual: Contrastive Features
Dhurandhar, et al. Explanations based on the missing: Towards contrastive explanations with pertinent negatives. NeurIPS 2018