Decision Trees Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

What are categorial and what are numeric attributes?

A

numeric (e.g., number of legs) and categorical (e.g., delicious/not delicious)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

classification trees and

regression trees difference?

A

cl: which produce categorical outputs

produce numeric outputs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is entropy?

A

Entropy represents the uncertainty of the data.
If all the data points belong to a single class, then there is no real uncertainty=low entropy. If the data points are evenly spread across the classes, there is a lot of uncertainty=high entropy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What do we have to take care of regarding entropy if we partition data?

A

partition needs low entropy if it splits the data into subsets that themselves have low entropy (i.e., are highly certain), and high entropy if it contains subsets that (are large and) have high entropy (i.e., are highly uncertain)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What do decision nodes and leaf nodes do?

A

decision nodes lead us through the tree and leaf nodes are the ends which give us predictions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What can random forests do?

A

They can aviod overfitting by building multiple decision trees and let them decide how to classify.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly