Week 11 - Explainability Flashcards

1
Q

Why do de need reasons for interpretability in ML models. Name 4 reasons

A
  1. Preventing Biases
  2. Verifying decisions
  3. Trust and acceptance
  4. Debugging
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is ‘explaining predictions’ in terms of ml

A

By explaining predictions we mean presenting textual or visual artifacts that provide a qualitative understanding of the relationship between the instance’s components and the model’s prediction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are interpretable models

A

Some models are inherently explainable. This allows for exploring why a model made a certain decision

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are black box models

A

Some ml are considered black box. it is difficult to investigate and understand why a certain ml decision has been reaches

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Give 3 example models that are not inherently explainable

A
  • Neural networks
  • Random forest
  • SVM kernels
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are model-specific approaches for interpretation

A

model specific interpretation tools are limited to specific model classes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are model agnostic approaches

A

Model-agnostic tools can be used on any mahine-learning model and are applied after the model has been trained. These agnositc moethods usually work by analyszing feature input and output pairs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are the 3 interpretable model properties

A
  1. Linear
  2. Monotonicity: the relationship between a feature and the target outcome always goes in the same direction over the entire range of the feature
  3. Interaction: automatically includes interactions between features to predict the target outcome.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How can we interpret a linear regression model?

A
  1. Numerical feauture.
  2. Categorical feature
  3. Intercept
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are 3 model-agnostic methods of interpretability

A
  1. Model flexibility
  2. Explanation flexibility
  3. Representation flexibility
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does LIME stand for

A

Local Interpretable Model-Agnostic Explanations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What are surrogate models

A

Surrogate models are trained to approximate the predictions of the underlying black box model. Instead of training a global surrogate model, LIME focuses on training a local surrogate models to explain individual predictions. A surrogate model in machine learning is an approximate model used to mimic the behavior of a more complex or computationally expensive model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is Local fidelity

A

Fidelity shows how faithful the interpretation is to the while model. With non-linear models, it might be impossible to find an explanation that is faithfil unless it is the complete desciption of the model itself.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the difference between local fidelity and global fidelity

A

Local Fidelity: The accuracy of a surrogate model in approximating the true model’s behavior in a specific, localized region of the input space.

Global Fidelity: The accuracy of a surrogate model in approximating the true model’s behavior across the entire input space.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

The idea of LIME is to come up with a balanced trade-off between ____ and ____

A

interpretability and local fidelity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the LIME approach. Give steps

A
  • we have a black box model where you can input data points and get predictions of the model.
  • We can probe the box as often as you want to understand why the machine learning model made a certain prediction. LIME tests what happens to the predictions when you give variations of your data(perturbation) into the machine learning model.
  • LIME generates a new dataset
  • LIME then trains an interpretable model
17
Q

What are perturbed samples.

A

Data points that have been intentionally modified or slightly altered from their original values to explore the sensitivity of a model or to enhance robustness during training.

18
Q

What are the steps in LIME for interpreting image classification

A
  1. Modify the image by enabling and disambling some of the interpeatble components
  2. For each of the generated images the probability of a tree frog’s preseness on it is obtained and used to create a labeled data-set.
  3. Next we train a lienar model weighted on the created data-set
  4. The components with the highest weights are presented