Week 11 - Explainability Flashcards
Why do de need reasons for interpretability in ML models. Name 4 reasons
- Preventing Biases
- Verifying decisions
- Trust and acceptance
- Debugging
What is ‘explaining predictions’ in terms of ml
By explaining predictions we mean presenting textual or visual artifacts that provide a qualitative understanding of the relationship between the instance’s components and the model’s prediction
What are interpretable models
Some models are inherently explainable. This allows for exploring why a model made a certain decision
What are black box models
Some ml are considered black box. it is difficult to investigate and understand why a certain ml decision has been reaches
Give 3 example models that are not inherently explainable
- Neural networks
- Random forest
- SVM kernels
What are model-specific approaches for interpretation
model specific interpretation tools are limited to specific model classes.
What are model agnostic approaches
Model-agnostic tools can be used on any mahine-learning model and are applied after the model has been trained. These agnositc moethods usually work by analyszing feature input and output pairs.
What are the 3 interpretable model properties
- Linear
- Monotonicity: the relationship between a feature and the target outcome always goes in the same direction over the entire range of the feature
- Interaction: automatically includes interactions between features to predict the target outcome.
How can we interpret a linear regression model?
- Numerical feauture.
- Categorical feature
- Intercept
What are 3 model-agnostic methods of interpretability
- Model flexibility
- Explanation flexibility
- Representation flexibility
What does LIME stand for
Local Interpretable Model-Agnostic Explanations
What are surrogate models
Surrogate models are trained to approximate the predictions of the underlying black box model. Instead of training a global surrogate model, LIME focuses on training a local surrogate models to explain individual predictions. A surrogate model in machine learning is an approximate model used to mimic the behavior of a more complex or computationally expensive model
What is Local fidelity
Fidelity shows how faithful the interpretation is to the while model. With non-linear models, it might be impossible to find an explanation that is faithfil unless it is the complete desciption of the model itself.
What is the difference between local fidelity and global fidelity
Local Fidelity: The accuracy of a surrogate model in approximating the true model’s behavior in a specific, localized region of the input space.
Global Fidelity: The accuracy of a surrogate model in approximating the true model’s behavior across the entire input space.
The idea of LIME is to come up with a balanced trade-off between ____ and ____
interpretability and local fidelity