Lecture 13 - Explainable ML Flashcards

Introduction and interpretable models

1
Q

What is interpretability?

A

Ability to explain or to present in understandable terms to a human:

  • The degree to which a human can understand the cause of a decision
  • The degree to which a human can consistently predict the result of a model
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is an explanation?

A

Answer to a WHY question.

  • An explanation usually relates the feature values of an instance to its model prediction in a
    humanly understandable way.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Why do we need XAI?

A

Scientific Understanding,
Bias / fairness issues,
Model debugging and auditing,
Human-AI cooperation / acceptance,
Regulatory compliance,
High-risk applications & regulated industries.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is Intrinsic Interpretability?

A

A model is inherently easy to understand due to its simplicity, without requiring additional tools.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is Post-hoc Interpretability?

A

Another method or model is used after training to interpret the predictions of a complex or opaque model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is Model-Agnostic?

A

Interpretation methods that can be applied to any model, treating it as a black-box.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is Model-Specific?

A

Interpretation methods that require access to and make use of a model’s internal structure or components.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What does Local (Instance-Level) explains?

A

Explains individual predictions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does Global (Model-Level) explains?

A

Explains the overall model behavior.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What does Intermediate (Group-Level) explains?

A

Explains predictions for groups or subsets of data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does “Result of the interpretation method,” refer to?

A

refers to the form of explanation produced by XAI techniques.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Name five result of the interpretation method

A
  1. Result of the interpretation method
  2. Feature summary
    visualization
  3. Model internals
  4. Data points
  5. Global or local surrogates via intrinsically interpretable models
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the result of interpretation methods that provide a feature summary statistic?

A
  • Feature importance (E[feature importance]).
  • Feature interaction strengths.
    These summarize the influence or relationships between features across the model.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What does a feature summary visualization show?

A
  • Partial dependence plot (PDP): Shows the marginal effect of a feature on the predicted outcome.
  • Feature importance plot: Ranks features based on their influence on the model’s predictions.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is provided by interpretation methods that show model internals?

A
  • Linear model weights: Reflect the importance of features in linear models.
  • Decision tree structure: Provides “if-then” rules and decision paths.
  • Filters: Commonly found in neural networks, where filters capture feature patterns (e.g., in convolutional neural networks).
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the result of interpretation methods focused on data points?

A
  • Exemplars: Representative examples of data points that explain model behavior.
  • Counterfactual explanations: Show how changing certain features would change the model’s prediction.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What is the purpose of global or local surrogates?

A

Surrogates are interpretable models (like decision trees or linear models) that approximate the behavior of a complex “black-box” model, providing explanations at a global (whole model) or local (individual instance) level.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What does Expressive Power refer to in explanation methods?

A

Expressive Power refers to the “language” or structure of the explanations provided.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What does Translucency describe in the context of explanation methods?

A

Translucency describes how much the explanation method relies on looking into the machine learning model itself.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

What does Portability describe in explanation methods?

A

Portability refers to the range of machine learning models that an explanation method can be used with.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What does Algorithmic Complexity measure in explanation methods?

A

Algorithmic Complexity refers to the computational complexity of the explanation method, indicating how much time and resources it takes to compute the explanations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

What does Accuracy refer to in individual explanations?

A

Accuracy refers to how well an explanation can predict unseen data.

It measures whether the explanation can generalize beyond the examples used to generate it.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

What is Fidelity in the context of individual explanations?

A

Fidelity measures how well the explanation approximates the predictions of the black-box model.

High fidelity means the explanation closely mirrors the behavior of the original model

24
Q

What does Certainty/Confidence indicate in an explanation?

A

Certainty or confidence reflects whether the explanation captures the certainty of the machine learning model.
it should indicate how confident the model is in its prediction, particularly in uncertain or borderline cases.

25
Q

What does Comprehensibility/Plausibility refer to in explanations?

A

Comprehensibility or plausibility refers to how well humans can understand the explanation and how convincing or trust-building it is.

26
Q

What are some ways to measure Comprehensibility/Plausibility?

A
  1. User studies or surveys: Ask humans to evaluate how understandable or believable they find the explanations.
  2. task performance: Measure how well humans can make decisions or predictions based on the explanations.
  3. trust metrics: Analyze how much confidence users gain in the system after seeing explanations.
27
Q

How is Fidelity measured in explainability?

A

Fidelity measures how closely an explanation matches the predictions of the black-box model.
- it should be measured objectively by comparing the explanation’s output with the actual model’s predictions.

28
Q

How is Plausibility is measured in explainability?

A

Through user studies where participants evaluate whether they find the explanation credible.

29
Q

How is Comprehensibility evaluated in explainability?

A

It requires a user study to assess whether people can interpret and make sense of the explanation.

30
Q

What is simulatability, and how is it measured?

A

Simulatability measures the degree to which a human can predict or calculate the model’s outcome given the explanation.

It is assessed by testing whether humans can reproduce the mode’s predoctions based solely on the explanation provided, without needing access to the model itself.

31
Q

What does it mean for an explanation to be contrastive?

A

A contrastive explanation requires a point of reference for comparison.
It answers the question “why this and not that”

32
Q

What does it mean for an explanation to be selective?

A

A selective explanation is precise, focusing on a small set of the most important factors.

Humans can process only around 7±2 cognitive entities at a time, so good explanations are concise.

33
Q

Why is social context important for a good explanation?

A

A good explanation considers the social context of audience.

It tailors the explanation to the understanding, values, or expectations of the explainee.

34
Q

What does it mean for an explanation to be truthful?

A

A truthful explanation is scientifically sound and proves to be true in reality.

Good explanations must accurately reflect the underlying phenomena or model behavior.

35
Q

What makes an explanation General and Probable?

A

A good explanation should be general, meaning it can explain many events and has a high likelihood of being valid across different situations.

General causes are often considered good explanations because they apply broadly.

36
Q

Why should a good explanation be consistent with prior beliefs?

A

A good explanation is consistent with the prior beliefs if the explainee, making it easier for them to accept and understand.

It aligns with what the audience already knows or expects, making it more plausible and convincing.

37
Q

What are interpretable models?

A

Models that are transparent and simple enough to understand.

They stand for their own explanation and reflect perfect fidelity since the model is its own explanation.

38
Q

What are black-box models, and why do they require post-hoc explanations?

A

Black-box models are complex and opaque, making their internal processes difficult to understand.

They require post-hoc explanations because the model itself is not interpretable.

These explanations often cannot achieve perfect fidelity with the original model, meaning they may not fully reflect the model’s decision-making process.

39
Q

Why might the explanations for black-box models not make sense?

A

Post-hoc explanations may be oversimplified or inconsistent with the model’s actual processes.

40
Q

Why are black-box models often not suitable for combining external information with a risk assessment?

A

This is because they lack transparency and cannot easily incorporate or explain decisions involving external data.

41
Q

What are intrinsically interpretable models?

A

Intrinsically interpretable models are models that are transparent and easy to understand without needing post-hoc explanations.

Their structure and decision-making process can be understood by humans at a glance.

42
Q

What is a basic requirement for interpretability regarding features/input?

A

The features or input used by the model must be intelligible and humanly understandable.

This means the input data should be in a form that humans can easily interpret.

43
Q

What type of model structure is required for interpretability?

A

The model must be transparent and simple to understand.

44
Q

Why is a compact set of predictive feature important for interpretability?

A

A compact set of predictive features means the model uses a small number of key features to make prediction.

Fewer, more important features make the model easier to interpret and understand.

45
Q

What is a Linear Model?

A

A linear model predicts the target variable as a weighted sum of the input features.

The relationship between the features and the target is linear.

46
Q

What are Scoring Systems in the hierarchy of linear models?

A

Scoring systems are linear models where the weights are constrained to be integers.

47
Q

What is a Generalized Additive Model (GAM)?

A

A GAM is a model where the relationship between the input features and the target is additive but nonlinear.

It combines univariate component functions 𝑓𝑖 that can be nonlinear, but the effects are added together

48
Q

What is a link function in Generalized Additive Models?

A

A link function ℎ(.) connects the linear predictor (sum of the transformed features) to the expected value of the outcome variable.

It allows for the target variable to have a non-normal distribution.

49
Q

How do GAMs differ from linear models?

A

While linear models assume a strict linear relationship between features and the target, GAMs allow for nonlinear relationships through univariate component functions that transform individual features.

50
Q

What is the role of the component functions 𝑓𝑖 in a GAM?

A

component functions 𝑓𝑖 in a GAM are univariate functions that model the relationship between each feature 𝑥𝑖 and the target.

These functions can capture nonlinear patterns in the data for each individual feature.

51
Q

What is the key difference between Linear Models and Generalized Additive Models (GAMs)?

A

The key difference is that Linear Models assume a linear relationship between features and the target, while GAMs allow for nonlinear relationships by applying smooth functions to the features.

52
Q

What are Generalized Linear Models (GLMs)?

A

GLMs extend linear models by allowing for non-normal distributions of the response variable.

They use a link function to relate the linear predictor to the expected value of the outcome.

53
Q

How do Generalized Linear Models (GLMs) handle nonlinear relationships?

A

GLMs handle nonlinear relationships by creating new higher-order variables (e.g., polynomial terms) from the original features and applying a linear model in this new feature space.

54
Q

How is the higher-dimensional transformation in GLMs similar to kernel methods in SVM?

A

Both GLMs and kernel methods involve transforming the feature space to better capture complex relationships.

55
Q

What is the purpose of creating higher-order variables in GLMs?

A

Higher-order variables in GLMs allow the model to capture nonlinear relationships between features and the target, while still using a linear predictor in the transformed feature space.

56
Q

slide 7/29

A

nog maken