SHAP-XRT
The Shapley Value Meets
Conditional Independence Testing


Beepul Bharti, Jacopo Teneggi, Yaniv Romano, Jeremias Sulam
Explainable AI Seminars @ Imperial
5/30/2024
Beepul Bharti, Jacopo Teneggi, Yaniv Romano, Jeremias Sulam
Explainable AI Seminars @ Imperial
5/30/2024
[M. E. Kaminski, 2019]
E.U.: “right to an explanation” of decisions made
on individuals by algorithms
F.D.A.: “interpretability of the model outputs”
efficiency
nullity
symmetry
exponential complexity
Lloyd S Shapley. A value for n-person games. Contributions to the Theory of Games, 2(28):307–317, 1953.
Let be an - person cooperative game with characteristic function
How important is each player for the outcome of the game?
inputs
responses
predictor
inputs
responses
predictor
inputs
responses
predictor
What does it mean for a feature to receive a high Shapley Value?
Precise Notions of Importance
[Candes et al, 2018]
Do the features in S contain any additional information about the label Y that is not in the complement [n]∖S ?
Precise Notions of Importance
XRT: eXplanation Randomization Test
returns a valid p-value, p^i,S for the test above
Precise Notions of Importance
Given the Shapley coefficient of any feature
Then
and the (expected) p-value obtained for , i.e.
Theorem 1:
Teneggi, Bharti, Romano, and S. "SHAP-XRT: The Shapley Value Meets Conditional Independence Testing." TMLR (2023).
What does the Shapley value test?
Given the Shapley value for the i-th feature, and
Theorem 2:
Then, under Hglobal0, pglobal is a valid p-value and
Full Spectrum of Tests
S
S
p^4,S
p^2,S
Input features are not inherently
interpretable to users
So far
model
y^=f(x)
statistical importance of
xj
input features xj
→
concepts cj
input features xj
→
concepts cj
What are the important concepts for
the predictions of a model?
Global and local importance
Fixed predictor
Statistically-valid importance
Any set of concepts
x∈X
f: X→Rd
g: Rd→[0,1]k
c∈Rd×m
input
encoder
classifier
concepts
Z=c⊤H
semantics
Y^=g(f(X))
predictions
H=f(X)
embeddings
x∈X
f: X→Rd
g: Rd→[0,1]k
c∈Rd×m
input
encoder
classifier
concepts
Z=c⊤H
semantics
Y^=g(f(X))
predictions
H=f(X)
embeddings
FIXED
For a concept j∈[m]:={1,…,m}
type of dependence
marginal
over a population
global importance
H0,jG: Y^⊥⊥Zj
Zj
Y^
For a concept j∈[m]:={1,…,m}
type of dependence
conditional
over a population
global conditional importance
H0,jGC: Y^⊥⊥Zj∣Z−j
Zj
Y^
Z−j
For a concept j∈[m]:={1,…,m}
type of dependence
conditional
for a specific input
semantic XRT
H0,j,SLC: g(HS∪{j})=dg(HS),HC∼PH∣ZC=zC
zj
Y^
zS
Null hypothesis H0, significance level α∈(0,1)
Classical approach
collect data
analyze data
compute p-value
→
→
"Reject H0 if p≤α"
Sequential approach
Instantiate wealth process K0=1, Kt=Kt−1⋅(1+vtκt)
"Reject H0 when Kt≥1/α"
time
wealth
1/α
under H0
under H1
rejection time τ
Data-efficient
Only use the data is needed to reject
Adaptive
The harder to reject, the longer the test
↓
If concept c rejects faster than c′
then it is more important
Induces a natural rank of importance
All definitions of semantic importance are two-sample tests
H0: P=P
Global importance (SKIT)
H0,jG
PY^Zj=PY^×PZj
⟺
H0,jGC
Global conditional importance (c-SKIT)
⟺
PY^ZjZ−j=PY^Zj,Z−j, Zj∼PZj∣Z−j
H0,j,SLC
Local conditional importance (x-SKIT)
⟺
PY^S∪{j}=PY^S
All definitions of semantic importance are two-sample tests
H0: P=P
Global importance (SKIT)
H0,jG
PY^Zj=PY^×PZj
⟺
H0,jGC
Global conditional importance (c-SKIT)
⟺
PY^ZjZ−j=PY^Zj,Z−j, Zj∼PZj∣Z−j
H0,j,SLC
Local conditional importance (x-SKIT)
⟺
PY^S∪{j}=PY^S
novel procedures
global importance
global conditional importance
local conditional importance
Jeremias Sulam
Jacopo Teneggi
Yaniv Romano
Beepul Bharti
SHAP-XRT
IBYDMT