Week 8 Flashcards
What ML interpretation method separates the explanations from the machine learning model?
Model-agnostic interpretation methods
What is the advantage of using model-agnostic interpretation methods over model-specific ones?
Their flexibility. The same method can be used for any type of model.
What is the disadvantage of using only interpretable models instead of using model-agnostic interpretation methods?
Predictive performance is lost compared to other ML models, and you limit yourself to one type of model.
What are two alternatives to using model-agnostic interpretation methods?
- Use only interpretable models.
- Use model-specific interpretation methods.
What is the disadvantage of using model-specific interpretation methods compared to model-agnostic ones?
It binds you to one model type and it’s difficult to switch to something else.
Name three flexibilities that are desirable aspects of a model-agnostic explanation system:
- Model flexibility
- Explanation flexibility
- Representation flexibility
Model flexibility (as an aspect of a model-agnostic explanation system)
It can work with any ML model, such as random forests and deep neural networks.
Explanation flexibility (as an aspect of a model-agnostic explanation system)
It’s not limited to a certain form of explanation. For example, linear formula and graphics with feature importances are both options.
Representation flexibility (as an aspect of a model-agnostic explanation system)
It’s able to use a different feature representation as the model being explained.
How can we further distinguish model-agnostic interpretation methods?
Into local and global methods.
What do global model-agnostic interpretation methods describe?
How features affect the prediction on average.
What do local model-agnostic interpretation methods describe?
An individual prediction.
How are global model-agnostic methods often expressed?
As expected values based on the distribution of the data.
What is the partial dependence plot?
A feature effect plot: the expected prediction when all other features are marginalized out.
When are global interpretation methods particularly useful?
When you want to understand the general mechanisms in the data or debug a model (since they describe average behavior).
PDP (abbreviation)
Partial dependence plot
PD plot (abbreviation)
partial dependence plot
What does the PDP show?
The marginal effect one or two features have on the predicted outcome of a ML model. Can show whether relationship between target and a feature is linear, monotonic or more complex.
What does xs denote in the PD function for regression?
The features for which the PD function should be plotted.
What does XC denote in the PD function for regression?
The other features (so non-xs features) used in the ML model ^f.
How does PD work?
By marginalizing the ML model output over the distribution of the features in set C, so that the function shows the relationship between the features in set S we are interested in, and the predicted outcoe.
Give the PD function for regression in the form of an expectation:
EXC[^f(xS, XC)].
Give the PD function for regression in the form of an integral:
integral sign ^f(xS, XC) dP(XC).
How is the partial function ^fS estimated?
By calculating averages in the training data, using the Monte Carlo method.