Week 8 Flashcards
What ML interpretation method separates the explanations from the machine learning model?
Model-agnostic interpretation methods
What is the advantage of using model-agnostic interpretation methods over model-specific ones?
Their flexibility. The same method can be used for any type of model.
What is the disadvantage of using only interpretable models instead of using model-agnostic interpretation methods?
Predictive performance is lost compared to other ML models, and you limit yourself to one type of model.
What are two alternatives to using model-agnostic interpretation methods?
- Use only interpretable models.
- Use model-specific interpretation methods.
What is the disadvantage of using model-specific interpretation methods compared to model-agnostic ones?
It binds you to one model type and it’s difficult to switch to something else.
Name three flexibilities that are desirable aspects of a model-agnostic explanation system:
- Model flexibility
- Explanation flexibility
- Representation flexibility
Model flexibility (as an aspect of a model-agnostic explanation system)
It can work with any ML model, such as random forests and deep neural networks.
Explanation flexibility (as an aspect of a model-agnostic explanation system)
It’s not limited to a certain form of explanation. For example, linear formula and graphics with feature importances are both options.
Representation flexibility (as an aspect of a model-agnostic explanation system)
It’s able to use a different feature representation as the model being explained.
How can we further distinguish model-agnostic interpretation methods?
Into local and global methods.
What do global model-agnostic interpretation methods describe?
How features affect the prediction on average.
What do local model-agnostic interpretation methods describe?
An individual prediction.
How are global model-agnostic methods often expressed?
As expected values based on the distribution of the data.
What is the partial dependence plot?
A feature effect plot: the expected prediction when all other features are marginalized out.
When are global interpretation methods particularly useful?
When you want to understand the general mechanisms in the data or debug a model (since they describe average behavior).
PDP (abbreviation)
Partial dependence plot
PD plot (abbreviation)
partial dependence plot