SlideShare uma empresa Scribd logo
1 de 68
Baixar para ler offline
Machine Learning
Interpretability Inside Out
Mara Graziani, PhD student
What is interpretablity?
2
What is interpretablity?
3
Well… it depends.
“ Rising
barometer”
Desired Interpretability
4
Algorithm
icTransparency
Translatability
Decomposability
Explainability
Scope for interpretation
5
“Interpretability is defined as the ability to explain or
to present in understandable terms to a human*.”
* not all humans are familiar with Machine Learning
[Kim et al., 2018]
6
“The goal of interpretability is to describe the internals of
a system in a way that is understandable to humans*.”
* not all humans are familiar with Machine Learning
[Giplin et al., 2019]
7
[Wikipedia]
Interpretability as a human-centric “translation” problem
8
Explanation in the model
representation space
(input pixels, activations)
Explanation in the human
representation space
(a visualization, a concept, a
sentence, an important
factor)
[Kim et al., 2018]
Why do we need interpretablity?
9
Why do we need interpretablity?
10
Trained CNN
It’s a cat
Why do we need interpretablity?
11
Trained CNN
It’s a cat
YAY!
Why do we need interpretablity?
12
Trained CNN
It’s a cat
If you want to know more about networks easily fooled:
[Szegedy et al., 2013], [Nguyen et al., 2015],
[Papernot et al., 2016], [Moosavi-Dezfooli et al., 2017]
Why do we need interpretablity?
13
Trained CNN
It’s a cat
Oh…but…why?
If you want to know more about networks easily fooled:
[Szegedy et al., 2013], [Nguyen et al., 2015],
[Papernot et al., 2016], [Moosavi-Dezfooli et al., 2017]
Why do we need interpretablity?
14
Trained CNN
It’s a cat
…
If you want to know more about networks easily fooled:
[Szegedy et al., 2013], [Nguyen et al., 2015],
[Papernot et al., 2016], [Moosavi-Dezfooli et al., 2017]
Why questions
[Giplin et al., 2019]
Why is the model working?
Why is it not?
Why is the output like this?
Why is it not something else?
Why should we trust
the model?
Explain
Defend
actions
Gain trust …
develop better models!
Why do we need interpretablity?
15
Where do we need interpretablity?
16
HEALTH ROBOTICS ASSISTED DRIVING LAW SOCIAL SCIENCES
High-risk applications demanding also for accountability,
transparency, fairness, trust [FAccT conference]
FINANCE
Where do we need interpretability?
17
Where it is not needed
Why is it not needed?
One motivation does not cover it all …
privacy robustness
[Kim B, Hooker S.]
already well
studied problem
Where is it needed?
Why is it needed?
Safety, science, debugging.
aligning objectives
18
How do we achieve interpretablity?
19
Interpretability is challenging
and trending
Interpretability is challenging and trending
20
plot rights: B. Kim
Our goal today is the how
21
gain a clearer understanding
KNOWING WHAT TO
APPLY & WHERE
22
1. Inherently interpretable models
a. (Generalized) linear regression
b. Decision trees and rules
2. Interpreting complex models
a. From inside (opening the black box)
b. From outside (black-box)
3. What else? Use interpretability to develop
better models.
4. Q&A
Outline
Inside
Out
I will not talk about
dimensionality reduction
1. Inherently interpretable models
23 [Lou et al., 2012] [Caruana et al., 2015]
24
Linear Regression
Output is a weighted sum of each feature
A linear increase of a feature is translated into a proportional effected in the outcome
No interactions between features
25
Generalized Linear Regression
Family: normal
Link: identity
Family: binomial
Link: logit
Interpretation comes mostly from assumption on the data generation process
Complexity ~ generalization
We can replace with arbitrary distributions according to the data generation process
[Caruana et al. 2015]Generalized Additive Models
26
Decision trees and rules
Car’s model
Ah, it’s a Tesla
Autonomy >= 350?
Yes
Is it electric?
Yes
Trackable and explainable decisions.
Good for data interactions!
Only categorical features
Step functions…sharp!
Changes in data lead to
different tree
Complexity ~ depth [Kim B.]
27
Decision trees and rules
IF
PERSON CAPACITY < 2 &&
PRICE = ‘high’ &&
ELECTRIC = False &&
COMPANY LOGO = ‘horse’
THEN
car is a FERRARI
Intrinsic explanation
Sparse
Efficient
Not for regression
Only categorical features
Complexity ~ #rules
[Kim B.]
2. Interpret complex models
28
Deep Learning
Local Global
Model-specific Post-hoc
29
Helpful terminology
vs
= true for a specific instance
= true for an entire set of inputs
(ex. a class)
= model built-in analysis = applicable to any model
[Lipton, 2016]
InsideMMD [Kim et al., 2016].
IF [Koh et al., 2017]
Database search
Geometric approaches
Visualization
Deconv [Zeiler et al., 2013],
AM [Erhan et al., 2009],
Dissection [Bau et al., 2017]
SVCCA [Raghu et al., 2017]
Surrogates
2. Interpretability of deep learning inside out
30
Out Attribution…
To features
To concepts
LIME [Ribeiro et al., 2016].
SHAP [Lundberg et al., 2016]
Saliency [Simonyan et al., 2013],
CAM [Zhou et al., 2016],
LRP [Binder et al., 2016]
TCAV [Kim et al., 2018],
RCVs [Graziani et al., 2018]
Database search
31
What examples explain the data or the model?
Inside Global
Post-hoc
Database search
32
What examples explain the data or the model?
Prototype
= representative of all data
Criticism
= under-represented bit
Maximum Mean
Discrepancy
[Kim et al., 2016]
Global
Post-hoc
Influential Instances
33
Deleting one of these would strongly affect learning
Influence Functions
[Koh et al., 2017]
Best paper award!
Post-hoc
Global
Visualization
34
What is a neuron, a channel or a layer looking for?
Inside
35
What is a neuron, a channel or a layer looking for?
Deconvolution:
inverting convolution
operations
[Zeiler et al., 2013]
Figure credits: stanford CS230 (2018 Youtube)
Post-hoc
Local Deconvolutions
36
What is a neuron, a channel or a layer looking for?
Gradient Ascent
[Erhan et al., 2009],
[Olah et al. 2019]
This activation is maximized
Lucid toolbox
Post-hoc
Global Gradient Ascent
Network Dissection
37
What is a neuron, a channel or a layer looking for?
Network Dissection
[Bau et al., 2017]
~1K concepts:
Post-hoc
Global
Early training finds concepts, late training improves them
color, texture,
material, object, scene
Set of segmented
regions for all
concepts
38
Singular Vector Canonical Correlation Analysis
Can we compress what a layer has learned? [Raghu et al., 2017]
Responses of this
layer to all data
Singular Value Decomposition
& Canonical Correlation Analysis
Allows comparisons of layers, archiectures and insights on training dynamics
What did it look like and what can we
do here?
Post-hoc
Global
InsideMMD [Kim et al., 2016].
IF [Koh et al., 2017]
Database search
Geometric approaches
Visualization
Deconv [Zeiler et al., 2013],
AM [Erhan et al., 2009],
Dissection [Bau et al., 2017]
SVCCA [Raghu et al., 2017]
Surrogates
2. Interpretability of deep learning inside out
39
Out Attribution…
To features
To concepts
LIME [Ribeiro et al., 2016].
SHAP [Lundberg et al., 2016]
Saliency [Simonyan et al., 2013],
CAM [Zhou et al., 2016],
LRP [Binder et al., 2016]
TCAV [Kim et al., 2018],
RCVs [Graziani et al., 2018]
Surrogate models
40
Replacement is an interpretable model trained on the data and the black-box predictions
Complex decision function
Linear surrogate (R2 0.7)
Flexibility by different surrogates
Very approximative ….
Post-hocGlobal
Local Interpretable Model-agnostic Explanations (LIME)
41
Replacement is an interpretable model trained on the data and the black-box predictions
to explain each prediction individually
Local linear surrogate
Flexible, universal (post-hoc)
Size of local neighborhood undefined
[Ribeiro et al., 2017]
Local Interpretable Model-agnostic Explanations (LIME)
42
Replacement is an interpretable model trained on the data and the black-box predictions
to explain each prediction individually
Local linear surrogate
Flexible, universal (post-hoc)
Size of local neighborhood undefined
[Ribeiro et al., 2017]
Local Interpretable Model-agnostic Explanations (LIME)
43
Replacement is an interpretable model trained on the data and the black-box predictions
to explain each prediction individually
Local linear surrogate
Flexible, universal (post-hoc)
Size of local neighborhood undefined
[Ribeiro et al., 2017]Sampling of local instances not very robust
Local Interpretable Model-agnostic Explanations (LIME)
44
Replacement is an interpretable model trained on the data and the black-box predictions
to explain each prediction individually
Local linear surrogate
Flexible, universal (post-hoc)
Size of local neighborhood undefined
Sampling of local instances not very robust
Local Interpretable Model-agnostic Explanations (LIME)
45
Replacement is an interpretable model trained on the data and the black-box predictions
to explain each prediction individually
Local linear surrogate
Flexible, universal (post-hoc)
Size of local neighborhood undefined
[Ribeiro et al., 2017]Sampling of local instances not very robust
SHapley Additive exPlanations
46
A game theoretic approach of competing features
[Lundberg et al., 2017]
Attributes to each input feature the change in the expected model prediction when conditioning on that feature.
Unifying framework, direct for categorical features
All rights reserved. No reuse allowed without permission.
(which was not peer-reviewed) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity.
The copyright holder for this preprint.http://dx.doi.org/10.1101/206540doi:bioRxiv preprint first posted online Oct. 21, 2017;
Only qualitative on images, difficult abstraction
Attribution to features
47
Out
Saliency
48
Indicates which pixels need to be changed the least to affect the output the most.
∂output
∂input
Saliency**
49 Slide credits: Hooker S,
** Saliency was recently shown to be very unstable and failing to be reliable in edge cases such
as in randomized networks or compared to random attribution maps (see Remove and Retrain, ROAR)
Class Activation Mapping
50
Importance of the image
regions is given by the
projection of the weights
of the output layer on the
last layer’s convolutional
feature maps
[Zhou et al., 2015]
Greatly succesfull technique for its transparency and directness
Only qualitative evaluation
Little focus on multiple instances of the same object
Post-hocLocal
Attribution to concepts
51
Out
Concepts can define object categories, so what about their relevance in the classification?
Concept Activation Vectors
52
?
Vector of “striped” texture
∂output
∂vector
Directional
derivative
[Kim et al., 2018]
We collect examples of a
concept, i.e. “striped
texture”
We take the internal
activations (unrolled)
Classification of
“striped” vs
“random”
What about the relevance of “striped” texture in the classification of a zebra?
Generalized
saliency
Post-hocLocal Global
53
Out
What if the concept is non-binary?
Such as tumor extension, patient age, color, ..
Our research on
54
Out
Regression Concept Vectors
Continuous concepts are directly measured on the image
Regression Concept Vectors
55
Segmentation
(manual or
automatic)
Handcrafted
features, texture
descriptors, shape,
size, …
Take the internal
activations (aggregation)
Linear regression of
measures
∂output
∂vector
Directional
derivative
Generalized
saliency
[Graziani et al., 2018]
Best paper award, iMIMIC, MICCAI 2018!
Vector of “size”
Post-hocLocal Global
Application to health
56
1 Modeling of visual concepts
2 CNN explanation
Nuclei
Pleomorphism
Tubular formation
Mitotic count
Enlarged nuclei
Vesicular
appearance
Multiple nucleoli
Segmentation
size
Image
texture
descriptors
contrast
ASM
correlation
area
Nottingham
grading
guidelines
relevant for
positive class
image
tumor
probability
contrast
area
ASM
correlation
relevant for
negative class
black-box
state-of-art
model
contrast
ASM
correlation
area
high
low
Application to health
57
curvature mean
raw segmented
Individual relevancepn = 0.22
ppre = 0.70
pplus = 0.08
GT: normal; prediction: normal
cti median
cti meancurvature median
avg point diameter mean avg segment diameter median
raw segmented
pn = 0.99
ppre = 0.009
pplus = 0.0
1.082

1.168

0.118

0.447

5.24

-1 10
1.030

1.045

0.040

0.095

3.775

Retinopathy of prematurity
[Graziani et al., 2019]
[Yeche et al., 2019]
Radiomics
Image credits: Yeche et al. springer
Applications to computer vision
58
Understanding scale
Applications to computer vision
59
Image blue-ness
Color and texture were used to
reduce this loss!
[Graziani et al., 2019]
Interpreting intentionally flawed models
Our goal today is the how
60
gain a clearer understanding
WHAT TO
APPLY & WHERE?
61
WHAT TO APPLY & WHERE?
What do you need most?
In deep learning
Understand
each
component
User-friendly
explanations
Visual explanations on
single input
Global
understanding
Individually
Interactions
Make
comparisons
Surrogate
models
From dataset of
conceptual examples
Measuring attributes on
imagesRCVs CAVs
Gradient
Ascent
Geometry-
based
Before decision
layer
Class
Activation
Maps
Layerwise
Layerwise
Relevance
62
What else?
Current research
63
Can we use interpretability for
better control and development?
64
Can we use interpretability for better control and development?
It’s a cat
It has pointy ears…
And mustache!
Cats DO NOT have penguin legs….
Robustness to
adversarial
65
Can we use interpretability for better control and development?
Interpretability
analysis
Prior knowledge
User’s feedback
Additional targets for
our model output
Desired
features
Multi-task
Learning
Undesired
Features
Adversarial
Learning
66
TAKE AWAY
Pic credits: bannerengineering.com
Cartography to navigate
interpretability (slide 121)
Growing conferences, workshops:
Some interesting people and projects:
FAccT, Tutorial on Interpretable Machine Learning,
NeurIPS Interpretable ML, ICML Interpretable ML,
DL summer schools, AISTATS, CVPR, ICLR, ECCV,
ICCV, ECML, KDD, Workshop on Intepreting and
Explaining Visual AI Models, Tutorial on
Interpretable & transporent deep learning, WHI 2020
(virtual this year)… and many others!
B. Zhou (Torralba, MIT), B. Kim (Google Brain), G. Montavon (heatmapping.org), DARPA’s XAI, Ruth C.
Fong (Harvard & Microsoft), Finale Doshi-Velez (Harvard), A. Weller (Cambridge), S. Lundberg
(Microsoft), DARPA’s XAI Explainable Artificial Intelligence…and many others!
ML interpretability is human-centric, multi-faceted and should be tailored on a precise scope.
Thank you!
67
linear probing interpretationinternal  representation
classification [1] regression (ours)
TCAV [1]
Br (ours [3,4,5])
UBS [2]
post-hoc interpretability
Text
DL model
feedback
through
concepts
Modeling of prior knowledge
area
contrast
Update of objective function
Handcrafted ML features
ML interpretability for healthcare: our vision
mara.graziani@hevs.ch
@mormontre
@maragraziani
68
mara.graziani@hevs.ch
@mormontre
@maragraziani
Q&A?

Mais conteúdo relacionado

Mais procurados

Machine Learning Applications in NLP.ppt
Machine Learning Applications in NLP.pptMachine Learning Applications in NLP.ppt
Machine Learning Applications in NLP.pptbutest
 
Introduction to machine learning and model building using linear regression
Introduction to machine learning and model building using linear regressionIntroduction to machine learning and model building using linear regression
Introduction to machine learning and model building using linear regressionGirish Gore
 
Applied Artificial Intelligence Unit 3 Semester 3 MSc IT Part 2 Mumbai Univer...
Applied Artificial Intelligence Unit 3 Semester 3 MSc IT Part 2 Mumbai Univer...Applied Artificial Intelligence Unit 3 Semester 3 MSc IT Part 2 Mumbai Univer...
Applied Artificial Intelligence Unit 3 Semester 3 MSc IT Part 2 Mumbai Univer...Madhav Mishra
 
Factorization Machines with libFM
Factorization Machines with libFMFactorization Machines with libFM
Factorization Machines with libFMLiangjie Hong
 
Machine Learning Unit 1 Semester 3 MSc IT Part 2 Mumbai University
Machine Learning Unit 1 Semester 3  MSc IT Part 2 Mumbai UniversityMachine Learning Unit 1 Semester 3  MSc IT Part 2 Mumbai University
Machine Learning Unit 1 Semester 3 MSc IT Part 2 Mumbai UniversityMadhav Mishra
 
Been Kim - Interpretable machine learning, Nov 2015
Been Kim - Interpretable machine learning, Nov 2015Been Kim - Interpretable machine learning, Nov 2015
Been Kim - Interpretable machine learning, Nov 2015Seattle DAML meetup
 
CounterFactual Explanations.pdf
CounterFactual Explanations.pdfCounterFactual Explanations.pdf
CounterFactual Explanations.pdfBong-Ho Lee
 
DagdelenSiriwardaneY..
DagdelenSiriwardaneY..DagdelenSiriwardaneY..
DagdelenSiriwardaneY..butest
 
Introduction to Machine Learning
Introduction to Machine LearningIntroduction to Machine Learning
Introduction to Machine LearningShahar Cohen
 
Explainable AI - making ML and DL models more interpretable
Explainable AI - making ML and DL models more interpretableExplainable AI - making ML and DL models more interpretable
Explainable AI - making ML and DL models more interpretableAditya Bhattacharya
 
Deciphering AI - Unlocking the Black Box of AIML with State-of-the-Art Techno...
Deciphering AI - Unlocking the Black Box of AIML with State-of-the-Art Techno...Deciphering AI - Unlocking the Black Box of AIML with State-of-the-Art Techno...
Deciphering AI - Unlocking the Black Box of AIML with State-of-the-Art Techno...Analytics India Magazine
 
Basics of Machine Learning
Basics of Machine LearningBasics of Machine Learning
Basics of Machine Learningbutest
 
Introduction to Machine Learning
Introduction to Machine LearningIntroduction to Machine Learning
Introduction to Machine LearningRahul Jain
 
Applied Artificial Intelligence Unit 5 Semester 3 MSc IT Part 2 Mumbai Univer...
Applied Artificial Intelligence Unit 5 Semester 3 MSc IT Part 2 Mumbai Univer...Applied Artificial Intelligence Unit 5 Semester 3 MSc IT Part 2 Mumbai Univer...
Applied Artificial Intelligence Unit 5 Semester 3 MSc IT Part 2 Mumbai Univer...Madhav Mishra
 
Machine Learning in NLP
Machine Learning in NLPMachine Learning in NLP
Machine Learning in NLPVijay Ganti
 
Machine Learning for Dummies (without mathematics)
Machine Learning for Dummies (without mathematics)Machine Learning for Dummies (without mathematics)
Machine Learning for Dummies (without mathematics)ActiveEon
 
Machine Learning: A Fast Review
Machine Learning: A Fast ReviewMachine Learning: A Fast Review
Machine Learning: A Fast ReviewAhmad Ali Abin
 
Machine Learning: Generative and Discriminative Models
Machine Learning: Generative and Discriminative ModelsMachine Learning: Generative and Discriminative Models
Machine Learning: Generative and Discriminative Modelsbutest
 
Nautral Langauge Processing - Basics / Non Technical
Nautral Langauge Processing - Basics / Non Technical Nautral Langauge Processing - Basics / Non Technical
Nautral Langauge Processing - Basics / Non Technical Dhruv Gohil
 

Mais procurados (20)

Machine Learning Applications in NLP.ppt
Machine Learning Applications in NLP.pptMachine Learning Applications in NLP.ppt
Machine Learning Applications in NLP.ppt
 
Introduction to machine learning and model building using linear regression
Introduction to machine learning and model building using linear regressionIntroduction to machine learning and model building using linear regression
Introduction to machine learning and model building using linear regression
 
Applied Artificial Intelligence Unit 3 Semester 3 MSc IT Part 2 Mumbai Univer...
Applied Artificial Intelligence Unit 3 Semester 3 MSc IT Part 2 Mumbai Univer...Applied Artificial Intelligence Unit 3 Semester 3 MSc IT Part 2 Mumbai Univer...
Applied Artificial Intelligence Unit 3 Semester 3 MSc IT Part 2 Mumbai Univer...
 
Factorization Machines with libFM
Factorization Machines with libFMFactorization Machines with libFM
Factorization Machines with libFM
 
Machine Learning Unit 1 Semester 3 MSc IT Part 2 Mumbai University
Machine Learning Unit 1 Semester 3  MSc IT Part 2 Mumbai UniversityMachine Learning Unit 1 Semester 3  MSc IT Part 2 Mumbai University
Machine Learning Unit 1 Semester 3 MSc IT Part 2 Mumbai University
 
Plug play language_models
Plug play language_modelsPlug play language_models
Plug play language_models
 
Been Kim - Interpretable machine learning, Nov 2015
Been Kim - Interpretable machine learning, Nov 2015Been Kim - Interpretable machine learning, Nov 2015
Been Kim - Interpretable machine learning, Nov 2015
 
CounterFactual Explanations.pdf
CounterFactual Explanations.pdfCounterFactual Explanations.pdf
CounterFactual Explanations.pdf
 
DagdelenSiriwardaneY..
DagdelenSiriwardaneY..DagdelenSiriwardaneY..
DagdelenSiriwardaneY..
 
Introduction to Machine Learning
Introduction to Machine LearningIntroduction to Machine Learning
Introduction to Machine Learning
 
Explainable AI - making ML and DL models more interpretable
Explainable AI - making ML and DL models more interpretableExplainable AI - making ML and DL models more interpretable
Explainable AI - making ML and DL models more interpretable
 
Deciphering AI - Unlocking the Black Box of AIML with State-of-the-Art Techno...
Deciphering AI - Unlocking the Black Box of AIML with State-of-the-Art Techno...Deciphering AI - Unlocking the Black Box of AIML with State-of-the-Art Techno...
Deciphering AI - Unlocking the Black Box of AIML with State-of-the-Art Techno...
 
Basics of Machine Learning
Basics of Machine LearningBasics of Machine Learning
Basics of Machine Learning
 
Introduction to Machine Learning
Introduction to Machine LearningIntroduction to Machine Learning
Introduction to Machine Learning
 
Applied Artificial Intelligence Unit 5 Semester 3 MSc IT Part 2 Mumbai Univer...
Applied Artificial Intelligence Unit 5 Semester 3 MSc IT Part 2 Mumbai Univer...Applied Artificial Intelligence Unit 5 Semester 3 MSc IT Part 2 Mumbai Univer...
Applied Artificial Intelligence Unit 5 Semester 3 MSc IT Part 2 Mumbai Univer...
 
Machine Learning in NLP
Machine Learning in NLPMachine Learning in NLP
Machine Learning in NLP
 
Machine Learning for Dummies (without mathematics)
Machine Learning for Dummies (without mathematics)Machine Learning for Dummies (without mathematics)
Machine Learning for Dummies (without mathematics)
 
Machine Learning: A Fast Review
Machine Learning: A Fast ReviewMachine Learning: A Fast Review
Machine Learning: A Fast Review
 
Machine Learning: Generative and Discriminative Models
Machine Learning: Generative and Discriminative ModelsMachine Learning: Generative and Discriminative Models
Machine Learning: Generative and Discriminative Models
 
Nautral Langauge Processing - Basics / Non Technical
Nautral Langauge Processing - Basics / Non Technical Nautral Langauge Processing - Basics / Non Technical
Nautral Langauge Processing - Basics / Non Technical
 

Semelhante a ML Interpretability Inside Out

​​Explainability in AI and Recommender systems: let’s make it interactive!
​​Explainability in AI and Recommender systems: let’s make it interactive!​​Explainability in AI and Recommender systems: let’s make it interactive!
​​Explainability in AI and Recommender systems: let’s make it interactive!Eindhoven University of Technology / JADS
 
Personalized Retweet Prediction in Twitter
Personalized Retweet Prediction in TwitterPersonalized Retweet Prediction in Twitter
Personalized Retweet Prediction in TwitterLiangjie Hong
 
A Categorisation of Post-hoc Explanations for Predictive Models
A Categorisation of Post-hoc Explanations for Predictive ModelsA Categorisation of Post-hoc Explanations for Predictive Models
A Categorisation of Post-hoc Explanations for Predictive ModelsJane Dane
 
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for Everyone
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for EveryoneGDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for Everyone
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for EveryoneJames Anderson
 
Training at AI Frontiers 2018 - Ni Lao: Weakly Supervised Natural Language Un...
Training at AI Frontiers 2018 - Ni Lao: Weakly Supervised Natural Language Un...Training at AI Frontiers 2018 - Ni Lao: Weakly Supervised Natural Language Un...
Training at AI Frontiers 2018 - Ni Lao: Weakly Supervised Natural Language Un...AI Frontiers
 
Building AI Applications using Knowledge Graphs
Building AI Applications using Knowledge GraphsBuilding AI Applications using Knowledge Graphs
Building AI Applications using Knowledge GraphsAndre Freitas
 
Bridging the gap between AI and UI - DSI Vienna - full version
Bridging the gap between AI and UI - DSI Vienna - full versionBridging the gap between AI and UI - DSI Vienna - full version
Bridging the gap between AI and UI - DSI Vienna - full versionLiad Magen
 
LinkedUp kickoff meeting session 4
LinkedUp kickoff meeting session 4LinkedUp kickoff meeting session 4
LinkedUp kickoff meeting session 4Hendrik Drachsler
 
Landscape of AI/ML in 2023
Landscape of AI/ML in 2023Landscape of AI/ML in 2023
Landscape of AI/ML in 2023HyunJoon Jung
 
Feature Fusion and Classifier Ensemble Technique for Robust Face Recognition
Feature Fusion and Classifier Ensemble Technique for Robust Face RecognitionFeature Fusion and Classifier Ensemble Technique for Robust Face Recognition
Feature Fusion and Classifier Ensemble Technique for Robust Face RecognitionCSCJournals
 
Automating Software Development Using Artificial Intelligence (AI)
Automating Software Development Using Artificial Intelligence (AI)Automating Software Development Using Artificial Intelligence (AI)
Automating Software Development Using Artificial Intelligence (AI)Jeremy Bradbury
 
Artificial Intelligence power point presentation document
Artificial Intelligence power point presentation documentArtificial Intelligence power point presentation document
Artificial Intelligence power point presentation documentDavid Raj Kanthi
 
Incorporating Word Embeddings into Open Directory Project Based Large-Scale C...
Incorporating Word Embeddings into Open Directory Project Based Large-Scale C...Incorporating Word Embeddings into Open Directory Project Based Large-Scale C...
Incorporating Word Embeddings into Open Directory Project Based Large-Scale C...Korea University
 
GDG Community Day 2023 - Interpretable ML in production
GDG Community Day 2023 - Interpretable ML in productionGDG Community Day 2023 - Interpretable ML in production
GDG Community Day 2023 - Interpretable ML in productionSARADINDU SENGUPTA
 
THE IMPACT OF USING VISUAL PROGRAMMING ENVIRONMENT TOWARDS COLLEGE STUDENTS’ ...
THE IMPACT OF USING VISUAL PROGRAMMING ENVIRONMENT TOWARDS COLLEGE STUDENTS’ ...THE IMPACT OF USING VISUAL PROGRAMMING ENVIRONMENT TOWARDS COLLEGE STUDENTS’ ...
THE IMPACT OF USING VISUAL PROGRAMMING ENVIRONMENT TOWARDS COLLEGE STUDENTS’ ...ijma
 
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...Codiax
 
Deep Learning & NLP: Graphs to the Rescue!
Deep Learning & NLP: Graphs to the Rescue!Deep Learning & NLP: Graphs to the Rescue!
Deep Learning & NLP: Graphs to the Rescue!Roelof Pieters
 
Rsqrd AI: Recent Advances in Explainable Machine Learning Research
Rsqrd AI: Recent Advances in Explainable Machine Learning ResearchRsqrd AI: Recent Advances in Explainable Machine Learning Research
Rsqrd AI: Recent Advances in Explainable Machine Learning ResearchSanjana Chowdhury
 

Semelhante a ML Interpretability Inside Out (20)

graziani_bias.pdf
graziani_bias.pdfgraziani_bias.pdf
graziani_bias.pdf
 
​​Explainability in AI and Recommender systems: let’s make it interactive!
​​Explainability in AI and Recommender systems: let’s make it interactive!​​Explainability in AI and Recommender systems: let’s make it interactive!
​​Explainability in AI and Recommender systems: let’s make it interactive!
 
Personalized Retweet Prediction in Twitter
Personalized Retweet Prediction in TwitterPersonalized Retweet Prediction in Twitter
Personalized Retweet Prediction in Twitter
 
A Categorisation of Post-hoc Explanations for Predictive Models
A Categorisation of Post-hoc Explanations for Predictive ModelsA Categorisation of Post-hoc Explanations for Predictive Models
A Categorisation of Post-hoc Explanations for Predictive Models
 
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for Everyone
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for EveryoneGDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for Everyone
GDG Cloud Southlake #17: Meg Dickey-Kurdziolek: Explainable AI is for Everyone
 
Training at AI Frontiers 2018 - Ni Lao: Weakly Supervised Natural Language Un...
Training at AI Frontiers 2018 - Ni Lao: Weakly Supervised Natural Language Un...Training at AI Frontiers 2018 - Ni Lao: Weakly Supervised Natural Language Un...
Training at AI Frontiers 2018 - Ni Lao: Weakly Supervised Natural Language Un...
 
Building AI Applications using Knowledge Graphs
Building AI Applications using Knowledge GraphsBuilding AI Applications using Knowledge Graphs
Building AI Applications using Knowledge Graphs
 
Bridging the gap between AI and UI - DSI Vienna - full version
Bridging the gap between AI and UI - DSI Vienna - full versionBridging the gap between AI and UI - DSI Vienna - full version
Bridging the gap between AI and UI - DSI Vienna - full version
 
LinkedUp kickoff meeting session 4
LinkedUp kickoff meeting session 4LinkedUp kickoff meeting session 4
LinkedUp kickoff meeting session 4
 
Landscape of AI/ML in 2023
Landscape of AI/ML in 2023Landscape of AI/ML in 2023
Landscape of AI/ML in 2023
 
Feature Fusion and Classifier Ensemble Technique for Robust Face Recognition
Feature Fusion and Classifier Ensemble Technique for Robust Face RecognitionFeature Fusion and Classifier Ensemble Technique for Robust Face Recognition
Feature Fusion and Classifier Ensemble Technique for Robust Face Recognition
 
Automating Software Development Using Artificial Intelligence (AI)
Automating Software Development Using Artificial Intelligence (AI)Automating Software Development Using Artificial Intelligence (AI)
Automating Software Development Using Artificial Intelligence (AI)
 
Artificial Intelligence power point presentation document
Artificial Intelligence power point presentation documentArtificial Intelligence power point presentation document
Artificial Intelligence power point presentation document
 
Test for AI model
Test for AI modelTest for AI model
Test for AI model
 
Incorporating Word Embeddings into Open Directory Project Based Large-Scale C...
Incorporating Word Embeddings into Open Directory Project Based Large-Scale C...Incorporating Word Embeddings into Open Directory Project Based Large-Scale C...
Incorporating Word Embeddings into Open Directory Project Based Large-Scale C...
 
GDG Community Day 2023 - Interpretable ML in production
GDG Community Day 2023 - Interpretable ML in productionGDG Community Day 2023 - Interpretable ML in production
GDG Community Day 2023 - Interpretable ML in production
 
THE IMPACT OF USING VISUAL PROGRAMMING ENVIRONMENT TOWARDS COLLEGE STUDENTS’ ...
THE IMPACT OF USING VISUAL PROGRAMMING ENVIRONMENT TOWARDS COLLEGE STUDENTS’ ...THE IMPACT OF USING VISUAL PROGRAMMING ENVIRONMENT TOWARDS COLLEGE STUDENTS’ ...
THE IMPACT OF USING VISUAL PROGRAMMING ENVIRONMENT TOWARDS COLLEGE STUDENTS’ ...
 
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
 
Deep Learning & NLP: Graphs to the Rescue!
Deep Learning & NLP: Graphs to the Rescue!Deep Learning & NLP: Graphs to the Rescue!
Deep Learning & NLP: Graphs to the Rescue!
 
Rsqrd AI: Recent Advances in Explainable Machine Learning Research
Rsqrd AI: Recent Advances in Explainable Machine Learning ResearchRsqrd AI: Recent Advances in Explainable Machine Learning Research
Rsqrd AI: Recent Advances in Explainable Machine Learning Research
 

Último

GEAR TRAIN- BASIC CONCEPTS AND WORKING PRINCIPLE
GEAR TRAIN- BASIC CONCEPTS AND WORKING PRINCIPLEGEAR TRAIN- BASIC CONCEPTS AND WORKING PRINCIPLE
GEAR TRAIN- BASIC CONCEPTS AND WORKING PRINCIPLEselvakumar948
 
Computer Networks Basics of Network Devices
Computer Networks  Basics of Network DevicesComputer Networks  Basics of Network Devices
Computer Networks Basics of Network DevicesChandrakantDivate1
 
Unit 4_Part 1 CSE2001 Exception Handling and Function Template and Class Temp...
Unit 4_Part 1 CSE2001 Exception Handling and Function Template and Class Temp...Unit 4_Part 1 CSE2001 Exception Handling and Function Template and Class Temp...
Unit 4_Part 1 CSE2001 Exception Handling and Function Template and Class Temp...drmkjayanthikannan
 
Thermal Engineering-R & A / C - unit - V
Thermal Engineering-R & A / C - unit - VThermal Engineering-R & A / C - unit - V
Thermal Engineering-R & A / C - unit - VDineshKumar4165
 
Navigating Complexity: The Role of Trusted Partners and VIAS3D in Dassault Sy...
Navigating Complexity: The Role of Trusted Partners and VIAS3D in Dassault Sy...Navigating Complexity: The Role of Trusted Partners and VIAS3D in Dassault Sy...
Navigating Complexity: The Role of Trusted Partners and VIAS3D in Dassault Sy...Arindam Chakraborty, Ph.D., P.E. (CA, TX)
 
"Lesotho Leaps Forward: A Chronicle of Transformative Developments"
"Lesotho Leaps Forward: A Chronicle of Transformative Developments""Lesotho Leaps Forward: A Chronicle of Transformative Developments"
"Lesotho Leaps Forward: A Chronicle of Transformative Developments"mphochane1998
 
School management system project Report.pdf
School management system project Report.pdfSchool management system project Report.pdf
School management system project Report.pdfKamal Acharya
 
AIRCANVAS[1].pdf mini project for btech students
AIRCANVAS[1].pdf mini project for btech studentsAIRCANVAS[1].pdf mini project for btech students
AIRCANVAS[1].pdf mini project for btech studentsvanyagupta248
 
Block diagram reduction techniques in control systems.ppt
Block diagram reduction techniques in control systems.pptBlock diagram reduction techniques in control systems.ppt
Block diagram reduction techniques in control systems.pptNANDHAKUMARA10
 
Orlando’s Arnold Palmer Hospital Layout Strategy-1.pptx
Orlando’s Arnold Palmer Hospital Layout Strategy-1.pptxOrlando’s Arnold Palmer Hospital Layout Strategy-1.pptx
Orlando’s Arnold Palmer Hospital Layout Strategy-1.pptxMuhammadAsimMuhammad6
 
Thermal Engineering -unit - III & IV.ppt
Thermal Engineering -unit - III & IV.pptThermal Engineering -unit - III & IV.ppt
Thermal Engineering -unit - III & IV.pptDineshKumar4165
 
1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf
1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf
1_Introduction + EAM Vocabulary + how to navigate in EAM.pdfAldoGarca30
 
Moment Distribution Method For Btech Civil
Moment Distribution Method For Btech CivilMoment Distribution Method For Btech Civil
Moment Distribution Method For Btech CivilVinayVitekari
 
Design For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the startDesign For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the startQuintin Balsdon
 
DC MACHINE-Motoring and generation, Armature circuit equation
DC MACHINE-Motoring and generation, Armature circuit equationDC MACHINE-Motoring and generation, Armature circuit equation
DC MACHINE-Motoring and generation, Armature circuit equationBhangaleSonal
 
Online electricity billing project report..pdf
Online electricity billing project report..pdfOnline electricity billing project report..pdf
Online electricity billing project report..pdfKamal Acharya
 
Unleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leapUnleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leapRishantSharmaFr
 
PE 459 LECTURE 2- natural gas basic concepts and properties
PE 459 LECTURE 2- natural gas basic concepts and propertiesPE 459 LECTURE 2- natural gas basic concepts and properties
PE 459 LECTURE 2- natural gas basic concepts and propertiessarkmank1
 

Último (20)

GEAR TRAIN- BASIC CONCEPTS AND WORKING PRINCIPLE
GEAR TRAIN- BASIC CONCEPTS AND WORKING PRINCIPLEGEAR TRAIN- BASIC CONCEPTS AND WORKING PRINCIPLE
GEAR TRAIN- BASIC CONCEPTS AND WORKING PRINCIPLE
 
Computer Networks Basics of Network Devices
Computer Networks  Basics of Network DevicesComputer Networks  Basics of Network Devices
Computer Networks Basics of Network Devices
 
Unit 4_Part 1 CSE2001 Exception Handling and Function Template and Class Temp...
Unit 4_Part 1 CSE2001 Exception Handling and Function Template and Class Temp...Unit 4_Part 1 CSE2001 Exception Handling and Function Template and Class Temp...
Unit 4_Part 1 CSE2001 Exception Handling and Function Template and Class Temp...
 
Thermal Engineering-R & A / C - unit - V
Thermal Engineering-R & A / C - unit - VThermal Engineering-R & A / C - unit - V
Thermal Engineering-R & A / C - unit - V
 
Navigating Complexity: The Role of Trusted Partners and VIAS3D in Dassault Sy...
Navigating Complexity: The Role of Trusted Partners and VIAS3D in Dassault Sy...Navigating Complexity: The Role of Trusted Partners and VIAS3D in Dassault Sy...
Navigating Complexity: The Role of Trusted Partners and VIAS3D in Dassault Sy...
 
"Lesotho Leaps Forward: A Chronicle of Transformative Developments"
"Lesotho Leaps Forward: A Chronicle of Transformative Developments""Lesotho Leaps Forward: A Chronicle of Transformative Developments"
"Lesotho Leaps Forward: A Chronicle of Transformative Developments"
 
School management system project Report.pdf
School management system project Report.pdfSchool management system project Report.pdf
School management system project Report.pdf
 
FEA Based Level 3 Assessment of Deformed Tanks with Fluid Induced Loads
FEA Based Level 3 Assessment of Deformed Tanks with Fluid Induced LoadsFEA Based Level 3 Assessment of Deformed Tanks with Fluid Induced Loads
FEA Based Level 3 Assessment of Deformed Tanks with Fluid Induced Loads
 
AIRCANVAS[1].pdf mini project for btech students
AIRCANVAS[1].pdf mini project for btech studentsAIRCANVAS[1].pdf mini project for btech students
AIRCANVAS[1].pdf mini project for btech students
 
Block diagram reduction techniques in control systems.ppt
Block diagram reduction techniques in control systems.pptBlock diagram reduction techniques in control systems.ppt
Block diagram reduction techniques in control systems.ppt
 
Orlando’s Arnold Palmer Hospital Layout Strategy-1.pptx
Orlando’s Arnold Palmer Hospital Layout Strategy-1.pptxOrlando’s Arnold Palmer Hospital Layout Strategy-1.pptx
Orlando’s Arnold Palmer Hospital Layout Strategy-1.pptx
 
Thermal Engineering -unit - III & IV.ppt
Thermal Engineering -unit - III & IV.pptThermal Engineering -unit - III & IV.ppt
Thermal Engineering -unit - III & IV.ppt
 
Cara Menggugurkan Sperma Yang Masuk Rahim Biyar Tidak Hamil
Cara Menggugurkan Sperma Yang Masuk Rahim Biyar Tidak HamilCara Menggugurkan Sperma Yang Masuk Rahim Biyar Tidak Hamil
Cara Menggugurkan Sperma Yang Masuk Rahim Biyar Tidak Hamil
 
1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf
1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf
1_Introduction + EAM Vocabulary + how to navigate in EAM.pdf
 
Moment Distribution Method For Btech Civil
Moment Distribution Method For Btech CivilMoment Distribution Method For Btech Civil
Moment Distribution Method For Btech Civil
 
Design For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the startDesign For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the start
 
DC MACHINE-Motoring and generation, Armature circuit equation
DC MACHINE-Motoring and generation, Armature circuit equationDC MACHINE-Motoring and generation, Armature circuit equation
DC MACHINE-Motoring and generation, Armature circuit equation
 
Online electricity billing project report..pdf
Online electricity billing project report..pdfOnline electricity billing project report..pdf
Online electricity billing project report..pdf
 
Unleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leapUnleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leap
 
PE 459 LECTURE 2- natural gas basic concepts and properties
PE 459 LECTURE 2- natural gas basic concepts and propertiesPE 459 LECTURE 2- natural gas basic concepts and properties
PE 459 LECTURE 2- natural gas basic concepts and properties
 

ML Interpretability Inside Out

  • 1. Machine Learning Interpretability Inside Out Mara Graziani, PhD student
  • 3. What is interpretablity? 3 Well… it depends. “ Rising barometer”
  • 5. 5 “Interpretability is defined as the ability to explain or to present in understandable terms to a human*.” * not all humans are familiar with Machine Learning [Kim et al., 2018]
  • 6. 6 “The goal of interpretability is to describe the internals of a system in a way that is understandable to humans*.” * not all humans are familiar with Machine Learning [Giplin et al., 2019]
  • 8. Interpretability as a human-centric “translation” problem 8 Explanation in the model representation space (input pixels, activations) Explanation in the human representation space (a visualization, a concept, a sentence, an important factor) [Kim et al., 2018]
  • 9. Why do we need interpretablity? 9
  • 10. Why do we need interpretablity? 10 Trained CNN It’s a cat
  • 11. Why do we need interpretablity? 11 Trained CNN It’s a cat YAY!
  • 12. Why do we need interpretablity? 12 Trained CNN It’s a cat If you want to know more about networks easily fooled: [Szegedy et al., 2013], [Nguyen et al., 2015], [Papernot et al., 2016], [Moosavi-Dezfooli et al., 2017]
  • 13. Why do we need interpretablity? 13 Trained CNN It’s a cat Oh…but…why? If you want to know more about networks easily fooled: [Szegedy et al., 2013], [Nguyen et al., 2015], [Papernot et al., 2016], [Moosavi-Dezfooli et al., 2017]
  • 14. Why do we need interpretablity? 14 Trained CNN It’s a cat … If you want to know more about networks easily fooled: [Szegedy et al., 2013], [Nguyen et al., 2015], [Papernot et al., 2016], [Moosavi-Dezfooli et al., 2017]
  • 15. Why questions [Giplin et al., 2019] Why is the model working? Why is it not? Why is the output like this? Why is it not something else? Why should we trust the model? Explain Defend actions Gain trust … develop better models! Why do we need interpretablity? 15
  • 16. Where do we need interpretablity? 16
  • 17. HEALTH ROBOTICS ASSISTED DRIVING LAW SOCIAL SCIENCES High-risk applications demanding also for accountability, transparency, fairness, trust [FAccT conference] FINANCE Where do we need interpretability? 17
  • 18. Where it is not needed Why is it not needed? One motivation does not cover it all … privacy robustness [Kim B, Hooker S.] already well studied problem Where is it needed? Why is it needed? Safety, science, debugging. aligning objectives 18
  • 19. How do we achieve interpretablity? 19 Interpretability is challenging and trending
  • 20. Interpretability is challenging and trending 20 plot rights: B. Kim
  • 21. Our goal today is the how 21 gain a clearer understanding KNOWING WHAT TO APPLY & WHERE
  • 22. 22 1. Inherently interpretable models a. (Generalized) linear regression b. Decision trees and rules 2. Interpreting complex models a. From inside (opening the black box) b. From outside (black-box) 3. What else? Use interpretability to develop better models. 4. Q&A Outline Inside Out I will not talk about dimensionality reduction
  • 23. 1. Inherently interpretable models 23 [Lou et al., 2012] [Caruana et al., 2015]
  • 24. 24 Linear Regression Output is a weighted sum of each feature A linear increase of a feature is translated into a proportional effected in the outcome No interactions between features
  • 25. 25 Generalized Linear Regression Family: normal Link: identity Family: binomial Link: logit Interpretation comes mostly from assumption on the data generation process Complexity ~ generalization We can replace with arbitrary distributions according to the data generation process [Caruana et al. 2015]Generalized Additive Models
  • 26. 26 Decision trees and rules Car’s model Ah, it’s a Tesla Autonomy >= 350? Yes Is it electric? Yes Trackable and explainable decisions. Good for data interactions! Only categorical features Step functions…sharp! Changes in data lead to different tree Complexity ~ depth [Kim B.]
  • 27. 27 Decision trees and rules IF PERSON CAPACITY < 2 && PRICE = ‘high’ && ELECTRIC = False && COMPANY LOGO = ‘horse’ THEN car is a FERRARI Intrinsic explanation Sparse Efficient Not for regression Only categorical features Complexity ~ #rules [Kim B.]
  • 28. 2. Interpret complex models 28 Deep Learning
  • 29. Local Global Model-specific Post-hoc 29 Helpful terminology vs = true for a specific instance = true for an entire set of inputs (ex. a class) = model built-in analysis = applicable to any model [Lipton, 2016]
  • 30. InsideMMD [Kim et al., 2016]. IF [Koh et al., 2017] Database search Geometric approaches Visualization Deconv [Zeiler et al., 2013], AM [Erhan et al., 2009], Dissection [Bau et al., 2017] SVCCA [Raghu et al., 2017] Surrogates 2. Interpretability of deep learning inside out 30 Out Attribution… To features To concepts LIME [Ribeiro et al., 2016]. SHAP [Lundberg et al., 2016] Saliency [Simonyan et al., 2013], CAM [Zhou et al., 2016], LRP [Binder et al., 2016] TCAV [Kim et al., 2018], RCVs [Graziani et al., 2018]
  • 31. Database search 31 What examples explain the data or the model? Inside Global Post-hoc
  • 32. Database search 32 What examples explain the data or the model? Prototype = representative of all data Criticism = under-represented bit Maximum Mean Discrepancy [Kim et al., 2016] Global Post-hoc
  • 33. Influential Instances 33 Deleting one of these would strongly affect learning Influence Functions [Koh et al., 2017] Best paper award! Post-hoc Global
  • 34. Visualization 34 What is a neuron, a channel or a layer looking for? Inside
  • 35. 35 What is a neuron, a channel or a layer looking for? Deconvolution: inverting convolution operations [Zeiler et al., 2013] Figure credits: stanford CS230 (2018 Youtube) Post-hoc Local Deconvolutions
  • 36. 36 What is a neuron, a channel or a layer looking for? Gradient Ascent [Erhan et al., 2009], [Olah et al. 2019] This activation is maximized Lucid toolbox Post-hoc Global Gradient Ascent
  • 37. Network Dissection 37 What is a neuron, a channel or a layer looking for? Network Dissection [Bau et al., 2017] ~1K concepts: Post-hoc Global Early training finds concepts, late training improves them color, texture, material, object, scene Set of segmented regions for all concepts
  • 38. 38 Singular Vector Canonical Correlation Analysis Can we compress what a layer has learned? [Raghu et al., 2017] Responses of this layer to all data Singular Value Decomposition & Canonical Correlation Analysis Allows comparisons of layers, archiectures and insights on training dynamics What did it look like and what can we do here? Post-hoc Global
  • 39. InsideMMD [Kim et al., 2016]. IF [Koh et al., 2017] Database search Geometric approaches Visualization Deconv [Zeiler et al., 2013], AM [Erhan et al., 2009], Dissection [Bau et al., 2017] SVCCA [Raghu et al., 2017] Surrogates 2. Interpretability of deep learning inside out 39 Out Attribution… To features To concepts LIME [Ribeiro et al., 2016]. SHAP [Lundberg et al., 2016] Saliency [Simonyan et al., 2013], CAM [Zhou et al., 2016], LRP [Binder et al., 2016] TCAV [Kim et al., 2018], RCVs [Graziani et al., 2018]
  • 40. Surrogate models 40 Replacement is an interpretable model trained on the data and the black-box predictions Complex decision function Linear surrogate (R2 0.7) Flexibility by different surrogates Very approximative …. Post-hocGlobal
  • 41. Local Interpretable Model-agnostic Explanations (LIME) 41 Replacement is an interpretable model trained on the data and the black-box predictions to explain each prediction individually Local linear surrogate Flexible, universal (post-hoc) Size of local neighborhood undefined [Ribeiro et al., 2017]
  • 42. Local Interpretable Model-agnostic Explanations (LIME) 42 Replacement is an interpretable model trained on the data and the black-box predictions to explain each prediction individually Local linear surrogate Flexible, universal (post-hoc) Size of local neighborhood undefined [Ribeiro et al., 2017]
  • 43. Local Interpretable Model-agnostic Explanations (LIME) 43 Replacement is an interpretable model trained on the data and the black-box predictions to explain each prediction individually Local linear surrogate Flexible, universal (post-hoc) Size of local neighborhood undefined [Ribeiro et al., 2017]Sampling of local instances not very robust
  • 44. Local Interpretable Model-agnostic Explanations (LIME) 44 Replacement is an interpretable model trained on the data and the black-box predictions to explain each prediction individually Local linear surrogate Flexible, universal (post-hoc) Size of local neighborhood undefined Sampling of local instances not very robust
  • 45. Local Interpretable Model-agnostic Explanations (LIME) 45 Replacement is an interpretable model trained on the data and the black-box predictions to explain each prediction individually Local linear surrogate Flexible, universal (post-hoc) Size of local neighborhood undefined [Ribeiro et al., 2017]Sampling of local instances not very robust
  • 46. SHapley Additive exPlanations 46 A game theoretic approach of competing features [Lundberg et al., 2017] Attributes to each input feature the change in the expected model prediction when conditioning on that feature. Unifying framework, direct for categorical features All rights reserved. No reuse allowed without permission. (which was not peer-reviewed) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. The copyright holder for this preprint.http://dx.doi.org/10.1101/206540doi:bioRxiv preprint first posted online Oct. 21, 2017; Only qualitative on images, difficult abstraction
  • 48. Saliency 48 Indicates which pixels need to be changed the least to affect the output the most. ∂output ∂input
  • 49. Saliency** 49 Slide credits: Hooker S, ** Saliency was recently shown to be very unstable and failing to be reliable in edge cases such as in randomized networks or compared to random attribution maps (see Remove and Retrain, ROAR)
  • 50. Class Activation Mapping 50 Importance of the image regions is given by the projection of the weights of the output layer on the last layer’s convolutional feature maps [Zhou et al., 2015] Greatly succesfull technique for its transparency and directness Only qualitative evaluation Little focus on multiple instances of the same object Post-hocLocal
  • 51. Attribution to concepts 51 Out Concepts can define object categories, so what about their relevance in the classification?
  • 52. Concept Activation Vectors 52 ? Vector of “striped” texture ∂output ∂vector Directional derivative [Kim et al., 2018] We collect examples of a concept, i.e. “striped texture” We take the internal activations (unrolled) Classification of “striped” vs “random” What about the relevance of “striped” texture in the classification of a zebra? Generalized saliency Post-hocLocal Global
  • 53. 53 Out What if the concept is non-binary? Such as tumor extension, patient age, color, ..
  • 54. Our research on 54 Out Regression Concept Vectors Continuous concepts are directly measured on the image
  • 55. Regression Concept Vectors 55 Segmentation (manual or automatic) Handcrafted features, texture descriptors, shape, size, … Take the internal activations (aggregation) Linear regression of measures ∂output ∂vector Directional derivative Generalized saliency [Graziani et al., 2018] Best paper award, iMIMIC, MICCAI 2018! Vector of “size” Post-hocLocal Global
  • 56. Application to health 56 1 Modeling of visual concepts 2 CNN explanation Nuclei Pleomorphism Tubular formation Mitotic count Enlarged nuclei Vesicular appearance Multiple nucleoli Segmentation size Image texture descriptors contrast ASM correlation area Nottingham grading guidelines relevant for positive class image tumor probability contrast area ASM correlation relevant for negative class black-box state-of-art model contrast ASM correlation area high low
  • 57. Application to health 57 curvature mean raw segmented Individual relevancepn = 0.22 ppre = 0.70 pplus = 0.08 GT: normal; prediction: normal cti median cti meancurvature median avg point diameter mean avg segment diameter median raw segmented pn = 0.99 ppre = 0.009 pplus = 0.0 1.082 1.168 0.118 0.447 5.24 -1 10 1.030 1.045 0.040 0.095 3.775 Retinopathy of prematurity [Graziani et al., 2019] [Yeche et al., 2019] Radiomics Image credits: Yeche et al. springer
  • 58. Applications to computer vision 58 Understanding scale
  • 59. Applications to computer vision 59 Image blue-ness Color and texture were used to reduce this loss! [Graziani et al., 2019] Interpreting intentionally flawed models
  • 60. Our goal today is the how 60 gain a clearer understanding WHAT TO APPLY & WHERE?
  • 61. 61 WHAT TO APPLY & WHERE? What do you need most? In deep learning Understand each component User-friendly explanations Visual explanations on single input Global understanding Individually Interactions Make comparisons Surrogate models From dataset of conceptual examples Measuring attributes on imagesRCVs CAVs Gradient Ascent Geometry- based Before decision layer Class Activation Maps Layerwise Layerwise Relevance
  • 63. Current research 63 Can we use interpretability for better control and development?
  • 64. 64 Can we use interpretability for better control and development? It’s a cat It has pointy ears… And mustache! Cats DO NOT have penguin legs…. Robustness to adversarial
  • 65. 65 Can we use interpretability for better control and development? Interpretability analysis Prior knowledge User’s feedback Additional targets for our model output Desired features Multi-task Learning Undesired Features Adversarial Learning
  • 66. 66 TAKE AWAY Pic credits: bannerengineering.com Cartography to navigate interpretability (slide 121) Growing conferences, workshops: Some interesting people and projects: FAccT, Tutorial on Interpretable Machine Learning, NeurIPS Interpretable ML, ICML Interpretable ML, DL summer schools, AISTATS, CVPR, ICLR, ECCV, ICCV, ECML, KDD, Workshop on Intepreting and Explaining Visual AI Models, Tutorial on Interpretable & transporent deep learning, WHI 2020 (virtual this year)… and many others! B. Zhou (Torralba, MIT), B. Kim (Google Brain), G. Montavon (heatmapping.org), DARPA’s XAI, Ruth C. Fong (Harvard & Microsoft), Finale Doshi-Velez (Harvard), A. Weller (Cambridge), S. Lundberg (Microsoft), DARPA’s XAI Explainable Artificial Intelligence…and many others! ML interpretability is human-centric, multi-faceted and should be tailored on a precise scope.
  • 67. Thank you! 67 linear probing interpretationinternal  representation classification [1] regression (ours) TCAV [1] Br (ours [3,4,5]) UBS [2] post-hoc interpretability Text DL model feedback through concepts Modeling of prior knowledge area contrast Update of objective function Handcrafted ML features ML interpretability for healthcare: our vision mara.graziani@hevs.ch @mormontre @maragraziani