decision trees Flashcards

1
Q

why do people use decision trees

A

simpliscity and interperability
extendability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what is the predictive model structure of a decision learning tree

A

root or internal node: a feature
leaf node : target value
branch : represents a decision rule

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

what does a root or internal node represent in a decision learning tree

A

root or internal node: a feature

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

what does a leaf node represent in a decision learning tree

A

leaf node : target value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what does a branch represent in a decision learning tree

A

branch : represents a decision rule

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what are the types of decision trees

A

classification : target variable takes categorical data
regression : target variable takes continuous data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what are the difference between classification and regression trees

A

classification : target variable takes categorical data
regression : target variable takes continuous data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

what are classification trees

A

DTs where target variable takes categorical data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what are regression trees

A

DTs where target variable takes continuous data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

whgat does information gain for DTs do

A

measures reduction in entropy after splitting a dataset based on a feature (not features)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

what is the equation for information gain

A

I(Y,X) = H(Y) - H(Y|X)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

what does each part of the equation stand for in information gain
I(Y,X) = H(Y) - H(Y|X)

A

Y - random variable representing target
X - random variable representing feature of input sample
H(Y) - entropy of y
H(X|Y) - conditional entropy of y given x

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

what does information gain do

A

quantifies improvement of classifying labels after using a feature to split the dataset
feature that maxmises information gain is chosen for the split

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q
A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly