Explainable ML

A brief overview with practical examples

Responsible ML

discriminative model

\(\rightarrow\)

input

prediction

Real-world

performance

Reproducibility

Explainability

Fairness

Privacy

\(\rightarrow\)

Responsible ML

discriminative model

\(\rightarrow\)

input

prediction

Real-world

performance

Reproducibility

Explainability

Fairness

Privacy

\(\rightarrow\)

The black-box problem

Explaining model predictions

\(=\)

"hemorrhage"

\(f\)

\((\)

\()\)

What are the important parts of

the image for this prediction?

\[\downarrow\]

A subset of features \(C\) with \(f(x_C) \approx f(x)\)

Local Explanation

Machine learning 🤝 linear models

Can we use the coefficients of a linear

model to explain its predictions?

Consider

\[f(x) = \beta_0 + \beta_1x_1 + \beta_2x_2 + \dots + \beta_dx_d\]

then

\[\beta_j = \frac{\partial f}{\partial x_j}\]

Machine learning 🤝 linear models

Definition (Conditional indepedence)

Let \(X, Y, Z\) be random variables

\[Y \perp \!\!\! \perp X \mid Z \iff p(y \mid x,z) = p(y \mid z)\]

For

\[y = f(x) = \beta_0 + \beta_1x_1 + \beta_2x_2 + \dots + \beta_dx_d\]

it holds that

\[\beta_j = 0 \iff Y \perp \!\!\! \perp X_j \mid X_{-j} \]

\[\downarrow\]

Linear models are inherently interpretable

Interpretable models vs explanations

Is there a tradeoff between inherently interpretable models and explanations?

Interpretable models

Explanation methods

- Gradients

LIME [Ribeiro, 16], CAM [Zhou, 16], GRAD-CAM [Selvaraju, 17]

- Information theory

RDE [Kolek, 22],  V-IP [Chattopadhyay, 23] 

- Game theory

SHAP [Lundberg, 17]

- Conditional independence

IRT [Burns, 20], XRT [T&B, 23]

- Semantic features

T-CAV [Kim, 18], PCBM [Yuksekgonul, 22]

🧐 What even is importance?

LIME [Ribeiro, 16]

TLDR;

We know linear models, so let's train one!

GRAD-CAM [Selvaraju, 17]

TLDR;

But automatic differentiation! Do a (fancy) backward pass

🧐 So why are we still talking about this?

Concerns on the validity of these methods [Aldebayo, 18]

🧐 So why are we still talking about this?

Concerns on the robustness of these methods [Ghorbani, 19]

🧐 So why are we still talking about this?

Provably negative results on gradient explanations [Shah, 21]

SHAP [Lundbger, 17]

\[\uparrow\]

How much did player \(1\) contribute

to the outcome of the game?

Consider a game with \(d\) players

\[\dots\]

\(1\)

\(2\)

\(d\)

SHAP [Lundbger, 17]

How much did player \(1\) contribute

to the outcome of the game?

\[\downarrow\]

\[\phi_j = \sum_{C \subseteq [d] \setminus \{j\}} w_C \left[v(x_{C \cup \{j\}}) - v(x_C)\right]\]

Definition (Shapley value [Shapley, 53])

For a cooperative game with \(d\) players and characteristic function \(v:~2^{[d]} \to \mathbb{R}^+\), the Shapley value of player \(j\) is

\(\phi_j\) is the average marginalized contribution of

player \(j\) to all possible subsets of players

SHAP [Lundbger, 17]

\[\phi_j = \sum_{C \subseteq [d] \setminus \{j\}} w_C~\mathbb{E} \left[f(\widetilde{X}_{C \cup \{j\}}) - f(\widetilde{X}_C)\right]\]

For machine learning models

and

where

\[\widetilde{X}_C = [x_C, X_{-C}],\quad X_{-C} \sim \mathcal{D}_{X_C = x_C}\]

game \(v\)

\(\rightarrow\)

model \(f\)

\(x\)

\(x_C\)

\(\widetilde{X}_C\)

SHAP [Lundbger, 17]

\[\phi_j = \sum_{C \subseteq [d] \setminus \{j\}} w_C~\mathbb{E} \left[f(\widetilde{X}_{C \cup \{j\}}) - f(\widetilde{X}_C)\right]\]

and

where

\[\widetilde{X}_C = [x_C, X_{-C}],\quad X_{-C} \sim \mathcal{D}_{X_C = x_C}\]

game \(v\)

\(\rightarrow\)

model \(f\)

Sound theoretical

properties

1. Subsets are exponential

2. Need to sample 

😍

😔

💡 Fast hierarchical games (h-Shap)

JT, Alexandre Luster, JS (2021) "Fast Hierarchical Games for Image Explanations", TPAMI

What do these problems

have in common?

sick cells

green lights

hemorrhage

Fast hierarchical games (h-Shap)

What do these problems

have in common?

\(\downarrow\)

If a part of the image is not important it cannot contain any important features

They all satisfy a Multiple Instance Learning assumption

\[f(x) = 1 \iff \exists C \subseteq [d]:~f(x_C) = 1\]

Fast hierarchical games (h-Shap)

If a part of the image is not important it cannot contain any important features

Fast hierarchical games (h-Shap)

If a part of the image is not important it cannot contain any important features

Fast hierarchical games (h-Shap)

If a part of the image is not important it cannot contain any important features

1. Runs in

\(O(k\log d)\)

 

2. Retrieves the correct Shapley values

Fast hierarchical games (h-Shap)

JT, Paul H Yi, JS (2023) "Examination-level Supervision for Deep Learning-based Intracranial Hemorrhage Detection on Head CT", Radiology: AI

 

Fast hierarchical games (h-Shap)

JT, Paul H Yi, JS (2023) "Examination-level Supervision for Deep Learning-based Intracranial Hemorrhage Detection on Head CT", Radiology: AI

 

Fast hierarchical games (h-Shap)

coding time!

🧐 What do large Shapley values mean?

Suppose

What do you report as

important, either or both?

\[\phi_1 = 0.25\]

\[\phi_2 = 0.35\]

and

\[\downarrow\]

Need to rigorously test for feature importance

💡 Conditional independence

Conditional independence null hypothesis

\[H_0:~Y \perp \!\!\! \perp X \mid Z\]

we can test for \(H_0\) with conditional randomization tests

reject \(H_0\) if \(\hat{p} \leq \alpha\)

\[\downarrow\]

\(p\)-values tell us the confidence we can

deem a feature important with

Definition (Type I error control)

If \(H_0\) is true, the probability of rejecting is below \(\alpha\) 

\[\mathbb{P}_{H_0}[\hat{p} \leq \alpha] \leq \alpha\]

💡 Conditional independence

Conditional independence null hypothesis

\[H_0:~Y \perp \!\!\! \perp X \mid Z\]

we can test for \(H_0\) with conditional randomization tests

reject \(H_0\) if \(\hat{p} \leq \alpha\)

\[\downarrow\]

Definition (Type I error control)

If \(H_0\) is true, the probability of rejecting is below \(\alpha\) 

\[\mathbb{P}_{H_0}[\hat{p} \leq \alpha] \leq \alpha\]

Can we test for a model's local

conditional independence?

Explanation Randomization Test (XRT)

JT*, Beepul Bharti*, Yaniv Romano, JS (2023) "SHAP-XRT: The Shapley Value Meets Conditional Independence Testing", TMLR

 

model

feature

subset

\(f\)

\(j \in [d]\)

\(C \subseteq [d] \setminus \{j\}\)

\(x\)

input

\[H^{\text{XRT}}_{0,j,C}:~f(\widetilde{X}_{C \cup \{j\}}) \overset{d}{=} f(\widetilde{X}_C)\]

\[\downarrow\]

Reject the null hypothesis if the observed value of \(x_j\) changes the distribution of the response of the model

Explanation Randomization Test (XRT)

Recall

\[\phi_j = \sum_{C \subseteq [d] \setminus \{j\}} w_C~\mathbb{E} \left[f(\widetilde{X}_{C \cup \{j\}}) - f(\widetilde{X}_C)\right]\]

turns out that

\[\gamma_{j,C}^2 \leq p^{\text{XRT}}_{j,C} \leq 1 - \gamma_{j,C}\]

Explanation Randomization Test (XRT)

Explanation Randomization Test (XRT)

Thank you!

Beepul Bharti

Alexandre Luster

Yaniv Romano

Jeremias Sulam

XRT

h-Shap

[04/17/2023] Explainable AI

By Jacopo Teneggi

[04/17/2023] Explainable AI

  • 30