ahmad.haj.mosa@pwc.com
LinkedIn: ahmad-haj-mosa
schneider.fabian@pwc.com
LinkedIn: fabian-schneider-24122379
source: DARPA
Explainability
Prediction Accuracy
Neural Nets
Deep
Learning
Statistical
Models
AOGs
SVMs
Graphical
Models
Bayesian
Belief Nets
SRL
MLNs
Markov
Models
Decision
Trees
Ensemble
Methods
Random
Forests
| System 1 | System 2 |
|---|---|
| drive a car on highways | drive a car in cities |
| come up with a good chess move (if you're a chess master) | point your attention towards the clowns at the circus |
| understands simple sentences | understands law clauses |
| correlation | causation |
| hard to explain | easy to explain |
source: Thinking fast and slow by Daniel Kahneman
thinking fast
thinking slow
consciousness prior
learning slow
learning fast
hard explanation
easy
explanation
Sym
ML
Sym
ML
ML
ML
Sym
Sym
Sym
Sym
Data
Data
Data
RE
RE
RE
Sym
Data
Data
ML
Sym
ML
RE
Sym
source: arXiv:1810.02338: Neural-Symbolic VQA: Disentangling Reasoning from Vision and Language Understanding
Sym
Sym
Sym
RE
ML
NTN: learning representation of entities and their relations in a KB unsupervised using tensor representation
NTN: learning representation of entities and their relations in a KB unsupervised using tensor representation
Data
ML
Sym
Sym
RE
source: arXiv:1812.07997: Explanatory Graphs for CNNs
Initial
Target
Truth Table
Graph Representation
Shared Layer
Value Layer
Policy Layer
Bi-LSTM
Bi-LSTM
Atten-LSTM
Flatten
Dense
Dense
Bi-LSTM
Shared Layer
Value Layer
Policy Layer
Bi-LSTM
Bi-LSTM
Atten-LSTM
Flatten
Dense
Dense
Concatenation
Value of the selection [-1,1]
Bi-LSTM
Bi-LSTM
| 0 |
| 1 |
| 0 |
| 0 |
| 1 |
| 0 |
| 1 |
| 0 |
| 0 |
| 0 |
| 0 |
| 0 |
| 0 |
| 0 |
| 0 |
Bi-LSTM
Bi-LSTM
Atten-LSTM
Flatten
Dense
Dense
Bi-LSTM
Concatenation
| 0 |
| 1 |
| 0 |
| 1 |
| 0 |
| 0 |
| 1 |
| 0 |
| 0 |
| 0 |
| 0 |
| 0 |
| 0 |
| 0 |
| 0 |
Concatenation
Bi-LSTM
Bi-LSTM
Bi-LSTM
Atten-LSTM
Flatten
Dense
Dense
Bi-LSTM
Concatenation
| 1 | 0 | 0 | 0 |
| 0 | 1 | 0 | 0 |
| 0 | 0 | 0 | 0 |
| 0 | 1 | 1 | 0 |
| 0 | 0 | 0 | 0 |
| 0 | 0 | 0 | 0 |
| 1 | 0 | 0 | 0 |
| 0 | 1 | 0 | 0 |
| 0 | 0 | 1 | 0 |
| 0 | 0 | 0 | 0 |
| 0 | 0 | 1 | 0 |
| 1 | 1 | 0 | 0 |
| 0 | 0 | 0 | 0 |
| 0 | 0 | 0 | 0 |
Binary Logic
Multivariate Logic
Bi-LSTM
Atten-LSTM
Concatenation
Value of the selection [-1,1]
N W
Q P
Accumulated value W = W + v
Number of visits N = N + 1
Mean value Q = W / N
Prior probability of an action P
Active nodes in the graph: G
Bi-LSTM
N W
Q P
Accumulated value W = W + v
Number of visits N = N + 1
Mean value Q = W / N
Prior probability of an action P
Repeate the follwoing steps for M time:
Generate Binary data
Generate Target Rule
Simulate the MCT:
Choose the action that maximises Q + U
Continue until the leaf node is reached
Rollout W, N, Q, P
Train the policy layer using policy gradient
Train the value layer by minimizing MSE between W and the predicted W
Active nodes in the graph: G
Shared Layer
Value Layer
Policy Layer
Bi-LSTM
Atten-LSTM
Dense
Dense
Variational Disentagled AE
Dense Classifier
TSNE
LIME
LIME
CGN
(Ship_to = AT) & (Material_Tax =1) & (IncoTerm =1)
& (VAT_ID =AT) --> domestic full tax
[ not ( A || C ) || ( C && not ( B && E ) ) , f7 ( f1 ( A ) ) , not ( f8 ( f7 ( f1 ( A ) ) ) ) , not ( A || B ) || implication( C, D ) , ... ]
(.-) [[a] -> [b]] -> ([[b]] -> [a] -> [c]) -> [a] -> [c] (.-) funcList newFunc inp = newFunc (map ($inp) funcList) inp -- where -- funcList = Previous Node Function List -- newFunc = Node Function -- inp = Sample Input
predicted rules
searching for best fitting rules
re-validate rules
- Accuracy is not enough to trust an AI
- Accuracy vs Explainability is not a trade-off
Look for solutions in weird places:
- Try Functional Programming & Category Theory
Trends in XAI:
- closing the gap between Symbolic AI and DL
- disentangled representation
- Object-Oriented-Representation
- Computational Graph Networks
+ Don't let your robot read legal texts ;)
Fabian Schneider
PoC-Engineer & Researcher
schneider.fabian@pwc.com
Ahmad Haj Mosa
AI Researcher
ahmad.haj.mosa@pwc.com