lecture 9 Flashcards

1
Q

What are decision trees used for?

A

Classifying data by recursively splitting it based on feature values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is an internal node in a decision tree?

A

A decision point that tests a feature to split data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a leaf node in a decision tree?

A

A terminal node that assigns a class label.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Why are decision trees prone to overfitting?

A

They can learn noise and irrelevant patterns in training data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the primary advantage of decision trees?

A

They are easy to interpret and visualize.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is a common method to regularize decision trees?

A

Limiting tree depth or pruning unnecessary branches.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the standard algorithm for training decision trees?

A

The ID3 (Iterative Dichotomiser 3) algorithm.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the greedy strategy in decision tree learning?

A

Choosing the best split at each step without backtracking.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How do decision trees handle missing values?

A

They can assign the most common value or split on other features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is entropy in decision trees?

A

A measure of uncertainty in a dataset.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is information gain?

A

The reduction in entropy after splitting on a feature.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the Gini index?

A

A measure of how pure a split is in decision trees.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is overfitting in decision trees?

A

When a tree is too complex and memorizes training data instead of generalizing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is pruning in decision trees?

A

Removing branches that do not improve generalization.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is an ensemble model?

A

A combination of multiple models to improve performance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is bagging in ensemble learning?

A

Training multiple models on different subsets of the data and averaging predictions.

17
Q

What is the key idea behind random forests?

A

Building multiple decision trees with random feature selection to improve generalization.

18
Q

What is boosting in ensemble learning?

A

A method that trains models sequentially, giving more weight to misclassified instances.

19
Q

What is a weak learner in boosting?

A

A model that performs slightly better than random chance.

20
Q

What is AdaBoost?

A

A boosting algorithm that combines weak learners to create a strong classifier.

21
Q

What is gradient boosting?

A

A boosting method that optimizes a loss function by adding models sequentially.

22
Q

How does boosting differ from bagging?

A

Boosting trains models sequentially, while bagging trains models independently.

23
Q

What is the role of decision stumps in boosting?

A

They serve as simple weak learners in boosting algorithms.

24
Q

Why are ensemble models often better than single models?

A

They reduce variance and improve generalization.

25
Q

What is out-of-bag error in bagging?

A

The error estimated on samples not used in training each model.

26
Q

How do random forests differ from standard decision trees?

A

They use multiple trees with different random feature subsets.

27
Q

What is the main disadvantage of ensemble methods?

A

They are less interpretable compared to single decision trees.

28
Q

What is feature importance in decision trees?

A

A measure of how much a feature contributes to making decisions.

29
Q

What is the takeaway from decision trees and ensemble learning?

A

Combining multiple decision trees using ensemble methods improves model robustness and accuracy.