Decision Trees Flashcards

1
Q

What role does Entropy play?

A

Controls how the DT splits the data. It’s the measure of impurity in a bunch of examples. Impurity being how uniform are the classes in the example set.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the formula for Entropy?

A

Entropy = Sum(i) { p(i) * log2(p(i)) }, where p(i) = fraction of examples in class i, and sum(i) sums over all classes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the entropy of all examples being same class?

A

0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is information gain?

A

entropy(parent) - [weighted average]*entropy(children)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How does the decision tree utilize information gain?

A

It maximizes information gain to determine the splits.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Give intuitive explanation for how to remember bias

A

I can train the model with all sorts of data but it’s bias towards it’s original behavior and doesn’t change

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Give intuitive explanation for how to remember variance

A

It cares so very much about the data it’s being trained on and will change it’s behavior to match it’s behavior to whatever data it sees

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are DT strengths and weaknesses?

A

Strengths: Easy to use, graphically interpretable (knowledge representation), can build bigger classifiers from them with ensemble methods
Weaknesses: Prone to overfitting especially with lots of features,

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Give an example of remembering xor logic gate

A

When someone asks do you want to go to the movie or bowling. usually they mean xor as in pick one or the other but not both and not neither

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Decision tree space - compare xor and or

A

xor - exponential space for nodes

or - linear as you add nodes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is Inductive Bias

A

The inductive bias of a learning algorithm is the set of assumptions that the learner uses to predict outputs given inputs that it has not encountered. A classical example of an inductive bias is Occam’s Razor, assuming that the simplest consistent hypothesis about the target function is actually the best.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is Preference Bias?

A

A preference bias is when a learning algorithm incompletely searches a complete hypothesis
space. It chooses which part of the hypothesis space to search. An example is decision trees

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is Representation Bias?

A

A representation bias completely searches and incomplete hypothesis space. It searches the
whole space, but it is a small incomplete space.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly