This meetup took place in Mountain View on January 24th, 2019.
Description:
With the effort and contributions from researchers and practitioners from academia and industry, Machine Learning Interpretation has become a young sub-field of ML. However, the norms around its definition and understanding is still in its infancy and there are numerous different approaches emerging rapidly. However, there seems to be a lack of a consistent explanation framework to evaluate and consistently benchmark different algorithms - evaluating against interpretation, completeness and consistency of the algorithms.
The idea with the gym is to provide a controlled interactive environment for all forms of Machine Learning algorithms, - initially focusing on supervised predictive modeling problems, to allow analysts and data-scientists to explore, debug and generate insightful understanding of the models by
1.Model Validation: Ways to explore and validate black box ML systems enabling model comparison both globally and locally - identifying biases in the training data through interpretation.
2.What-if Analysis: An interactive environment where communication can happen i.e. enable learning through interactions. User having the ability to conduct "What-If" analysis - effect of single or multiple features and their interactions
3.Model Debugging: Ways to analyze the misbehavior of the model by exploring counterfactual examples(adversarial examples and training)
4. Interpretable Models: Ability to build natively interpretable models - with the goal to simplify complex models to enable better understanding.
The central concept with MLI gym is to have an interactive environment where one could explore and simulate variations in the world(a world post a model is operationalized) beyond the defined model metrics point estimates - e.g. ROC-AUC, confusion matrix, RMSE, R2 score and others.
Speaker's Bio:
Pramit is a Lead Data Scientist/ at H2O.ai. His area of interests is building Statistical/Machine Learning models(Bayesian and Frequentist Modeling techniques) to help the business realize their data-driven goals.
Currently, he is exploring "Model Interpretation" as means to efficiently understand the true nature of predictive models to enable model robustness and security. He believes effective Model Inference coupled with Adversarial training could lead to building trustworthy models with known blind spots. He has started an open source project Skater: https://github.com/datascienceinc/Skater to solve the need for Model Inference(The project is still in its early stages of development but check it out, always eager for feedback)
5. Confidential5
Machine Learning workflow
Figure: Shows an improved ML workflow showing how MLI fits
into a typical predictive analytical setup
**Reference: Adapted from ”Learning from data - Professor Yaser Abu-Mostafa”
Model Interpretation
Learnings by model inference
7. Confidential7
Problems Addressed by Driverless AI
• Supervised Learning
– Regression
– Classification
– Binary
– Multinomial
• Tabular structured data
– Numeric
– Categorical
– Time/Date
– Text
– Missing values are allowed
• IDENTICALLY AND
INDEPENDENTLY DISTRIBUTED
(IID) ROWS
• TIME-SERIES
– Single time-series
– Grouped time-series
– Time-series problems with a gap
between training and testing to
account for time to deploy
• NLP
– Categorization
– Sentiment
8. Confidential8
What is Model Interpretation?
• Ability to explain and present a model in a way that is understandable to humans.
-- Finale Doshi-Velez and Been Kim. “Towards a rigorous science of interpretable machine learning.” arXiv
preprint. 2017. https://arxiv.org/pdf/1702.08608.pdf
• “Why did you do A” or ”Why DIDN’T you do B”, ”Why CAN’T you do C” ?
-- Gilpin, Leilani H., et al. "Explaining Explanations: An Approach to Evaluating Interpretability of Machine
Learning." arXiv preprint arXiv:1806.00069 (2018).
• A Model’s result is self descriptive and needs no further explanation; expressed in terms of inputs and
outputs. “Mapping an abstract concept(e.g. predicted class) into a domain that human can make sense
of“
-- Gregoire Montavon et al. “Methods for Interpreting and Understanding Deep Neural Networks”
9. Confidential9 Confidential9
“A collection of visual and/or interactive artifacts
that provide a user with sufficient description of
the model behavior to accurately perform tasks
like evaluation, trusting, predicting, or improving
the model.”
Assistant Professor Sameer Singh, University of
California, Irvine
10. Confidential10
Motives for Model Interpretation
1. Helps in defining hypothesis for the problem
2. Debugging and improving an ML system –
mismatched objectives.
3. Exploring and discovering latent or hidden feature
interactions (useful for feature engineering/selection
and resolving preconceptions).
4. Understanding model variability to avoid over-fitting.
5. Helps in model comparison.
6. Building domain knowledge about a particular use
case.
7. Bring transparency to decision making to enable
trust and safety – ML System is making sound
decisions.
Model Maker(Producer)
1. Ability to share the explanations to consumers of the
predictive model.
2. Explain the model/algorithm.
3. Explain the key features driving the KPI.
4. Verify and validate the accountability of ML learning
systems, e.g., causes for false positives in credit scoring,
insurance claim frauds(identifying spurious correlations
is easier for humans).
5. Identify blind spots to prevent adversarial attacks or fix
dataset errors.
6. Right to explanation: Comply with data protection law
and regulations, e.g., EU’s GDPR.
Model Breaker(Consumer)
11. Confidential11
How will it help?
Currently
ML use-case = data munging + data scientist expertise + model building with availability to computation
How can we scale this 10x?
ML use-case = data munging + data scientist expertise + model building with availability to computation
Driverless Machine Learning + Machine Learning Interpretation(MLI)
12. Confidential12
Scope of Interpretation
Global Interpretation
Being able to explain the conditional interaction
between dependent (response) variables and
independent (predictor or explanatory)
variables based on the complete dataset.
Helpful in explaining the context of the decision
classification.
Local Interpretation
Being able to explain the conditional
interaction between dependent (response)
variables and independent (predictor or
explanatory) variables for a single row or a
subset of rows. Helpful in identifying local
trends and intuitions.
Global Interpretation
Local Interpretation
15. Confidential15
Experiment Settings
TimeAccuracy
• Relative interpretability – higher
values favor more interpretable
models
• The higher the interpretability
setting, the lower the complexity
of the engineered features and
of the final model(s)
Interpretability
19. Confidential19
PDP/ICE
p(HouseAge, Avg. Occupants per household) vs
Avg. House Value: One can observe that once the
avg. occupancy > 2, houseAge does not seem to
have much of an effect on the avg. house value
Image credit: Patrick Hall, Navdeep Gill,
h2o.ai team
• Helps in understanding interaction impact of two independent features in a low dimensional space
visually
• Helps in understanding the average partial dependence of the target function f(Y|X )on
subsetof features by marginalizing over rest of the features (complement set of features)
21. Confidential21
Shapley values
• Explanation idea is borrowed from coalitional game theory
• Initial idea was proposed by Lloyd S. Shapley (1953)
-- Lloyd S. Shapley, Alvin E. Roth, et al. The Shapley value: Essays in honor of Lloyd S. Shapley. Cambridge University
Press, 1988. URL: http://www.library.fa.ru/files/Roth2.pdf.
• Idea:
– Feature’s contribution is computed by observing the difference between model’s initial predictions
– Followed by it’s average prediction post perturbing the feature space
– One has to be careful with features which are computationally dependent while perturbing
• Defined as
-- Erik Strumbelj and Igor Kononenko. An Efficient Explanation of Individual Classifications using Game Theory. Journal of
Machine Learning Research,
11(Jan):1–18, 2010. URL: http://www.jmlr.org/papers/volume11/strumbelj10a/strumbelj10a.pdf.
-- Scott M. Lundberg and Su-In Lee. A Unified Approach to Interpreting Model Predictions. URL: http://papers.nips.cc/paper/
7062-a-unified-approach-to-interpreting-model-predictions.pdf.
• Scope of Interpretation: Helps in computing Globally and locally faithful feature importance
• Flexible: Can be adopted in different forms - model agnostic or model specific approximations (Tree SHAP for tree ensemble
methods, Linear SHAP)
22. Confidential22 Confidential22
Surrogate Decision Trees
• If the original DNN learned decision function is denoted as 𝑔 and set of predictions, g(X)=Ỳ, then a surrogate tree(𝘩tree) can be
learned such 𝘩tree (X, Ỳ) ≅ g(X)
• The faithfulness of the decisions generated by SDT depends on how precisely the tree surrogates captures the decisions learned
by the original estimator(g).
-- Craven, M., & Shavlik, J. W. (1996). Extracting tree-structured representations of trained networks. In Advances in neural
information processing systems (pp. 24-30).
A B
23. Confidential23
• Capture Feature Interactions
• Partial dependence function could possibly be extended to capture
feature interactions by applying statistical tests
Exploring: Capturing interactions
Image: Showing H-Statistics scores for all possible pairs of features ordered in-terms of importance
28. Confidential28
Can we think?
An interactive playground to simulate, infer and evaluate
all forms of model- what-if(scenario analysis), identifying
counterfactuals to establish model robustness
Image reference: Rajeswara et.al (http://arxiv.org/abs/1703.02660)