ML/DL Evaluation Flashcards

1
Q

Why is it important to evaluate and validate ML/DL models?

A

To ensure the model meets the desired goals and performs well on unseen data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are two common methods for splitting data for validation?

A
  1. Holdout strategy
  2. K-Fold Cross-Validation
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the holdout strategy in model validation?

A

The dataset is split into training, validation, and test sets to ensure model generalization.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is a typical split ratio for holdout validation?

A

Training: 60%, Validation: 20%, Testing: 20%.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How can we detect overfitting or underfitting?

A

By comparing training and validation performance:
- Overfitting: High training accuracy but low validation accuracy.
- Underfitting: Poor performance on both training and validation sets.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are two solutions to overfitting?

A
  1. Early stopping
  2. Regularization (e.g., L2 regularization)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is K-Fold Cross-Validation?

A

A technique where the dataset is divided into K subsets, iterating between training and validation to ensure robust evaluation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Why use K-Fold Cross-Validation?

A

It reduces bias by ensuring each data point is used for training and validation multiple times.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Why is hyperparameter optimization important?

A

It helps improve model performance and generalization.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are two common hyperparameter tuning methods?

A
  1. Grid Search
  2. Random Search
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How does Grid Search work?

A

It exhaustively evaluates all possible hyperparameter combinations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What are the pros and cons of Grid Search?

A
  • Pros: Thorough and systematic.
  • Cons: Computationally expensive for large parameter spaces.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How does Random Search work?

A

It randomly samples hyperparameters from a specified range.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are the pros and cons of Random Search?

A
  • Pros: More efficient for high-dimensional spaces.
  • Cons: No guarantee of finding the absolute best combination.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What are the two main types of classification?

A
  1. Binary classification
  2. Multi-class classification
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is a confusion matrix?

A

A table that summarizes the performance of a classification model.

17
Q

What are the four key components of a confusion matrix?

A
  1. True Positives (TP)
  2. False Negatives (FN)
  3. False Positives (FP)
  4. True Negatives (TN)
18
Q

Why is accuracy not always a reliable metric?

A

Accuracy can be misleading for imbalanced datasets.

19
Q

How is precision calculated?

A

Precision = TP / (TP + FP)

20
Q

How is recall calculated?

A

Recall = TP / (TP + FN)

21
Q

What is the F1-score?

A

The harmonic mean of precision and recall:
F1 = 2 × (Precision × Recall) / (Precision + Recall)

22
Q

What is AUC (Area Under the ROC Curve)?

A

It measures the area under the ROC curve, evaluating classifier performance across all threshold values.

23
Q

What is the goal of a regression model?

A

To predict a continuous target variable based on input features.

24
Q

What are two common regression metrics?

A
  1. Mean Absolute Error (MAE)
  2. Mean Squared Error (MSE)
25
Q

How do MAE and MSE differ?

A

MSE penalizes larger errors more than MAE.

26
Q

What is the Silhouette Coefficient?

A

A measure of how similar a data point is to its own cluster compared to other clusters.

27
Q

When should you use Grid Search?

A

For small, well-defined hyperparameter spaces.

28
Q

When should you use Random Search?

A

For large or continuous hyperparameter spaces.

29
Q

What is an alternative method for hyperparameter tuning?

A

Bayesian Optimization for complex problems.