Lesson 2 Flashcards

1
Q

What is the exhastive search?

Decision trees

A

Trying every possible way to split the data set and constructing every possible tree

For every feature the tree grows exponentially

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does a high impurity mean?

A

This means that a node has a more equal distribution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How can you calculate the depth of a decision tree?

A

It is the maximum amount of questions you can ask to get an answer

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are 7 advantages of Decision trees

A
  • easy to understand
  • can be visualised
  • handles both numerical and categorical data
  • requires little data preparation
  • captures non-linear relationships
  • feature importance
  • fast to train and predict
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are 4 drawbacks of decision trees

A
  • Prone to overfitting
  • Sensitive to minor details
  • Unstable
  • Biased toward features with more levels
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is ensemble learning?

A

When ‘weak learner’ models are combined either sequential or parallel.

“The wisdom of the crowd”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is Bagging?

Decision trees

A

Using different training sets to find the best possible decision tree

Parallel ensemble learning technique

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is Random Forest?

A

Making random trees with the same training data and finding the answer by using the most chosen over all of the DTs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly