Topic 3: Decision Trees Flashcards

1
Q

What is the key metric used by the ID3 algorithm to select attributes for splitting?

A

Information Gain.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is entropy in the context of decision trees?

A

A measure of uncertainty or impurity in the dataset.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How does pruning help in decision tree learning?

A

By reducing overfitting and removing irrelevant parts of the tree.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is a decision tree?

A

A flowchart-like structure where internal nodes represent tests on attributes, branches represent outcomes, and leaf nodes represent class labels or decisions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How does a decision tree classify an instance?

A

By traversing from the root to a leaf node, following the path defined by the instance’s feature values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What type of problems can decision trees handle?

A

Both classification (discrete outcomes) and regression (continuous outcomes).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is information gain?

A

A measure of how much uncertainty in the dataset is reduced after splitting on a particular attribute.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are the advantages of decision trees?

A

Easy to interpret and visualize.
Handles both numerical and categorical data.
Requires little preprocessing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What are the disadvantages of decision trees?

A

Prone to overfitting without pruning.
Can be unstable due to small changes in the data.
Less effective for large datasets with many classes or noise.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is pruning in decision trees?

A

A process of reducing the size of the tree by removing sections that provide little value to avoid overfitting.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the difference between pre-pruning and post-pruning?

A

Pre-pruning stops the tree’s growth early, based on a condition (e.g., max depth).
Post-pruning removes branches after the tree is fully grown.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How can cross-validation be used in pruning?

A

To determine whether removing a branch improves performance on unseen data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What causes overfitting in decision trees?

A

The tree grows too complex and fits the training data perfectly, capturing noise and irrelevant patterns.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How are decision trees used in medical diagnosis?

A

To classify diseases based on symptoms or test results.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly