Chapter 10- Ensemble Models Flashcards

1
Q

what is the principle of committee decision?

A

individual predictions combined appropriately should have better overall accuracy on average than an individual committee member

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what three methods can be used to combine decisions?

A

averaging, voting, probablistic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

what is the wisdom of crowds?

A

different versions make errors in different ways. combining them cancels out the errors of the individuals

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

what is bagging?

A

generate different training sets, by sampling with replacement to build a committee of models

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

is bagging a sequential or parallel method?

A

parallel

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what are random forests?

A

an extension of bagging applied to decision trees.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what are the two methods of generating randomisation for random forests?

A

bootstrap, as in bagging

random selection of features are each split point

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

why do we force a random selection of feature at each split point in random forests?

A

to ensure it doesnt choose the best feature at every split.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what does the lack of pruning of random forest trees ensure

A

that the models aren’t too simple and therefore too similar to each other.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

is adaboost parallel or sequential?

A

sequential

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

describe the adaboost algorithm

A

a new model trained at each round. at the end of each round the mis-classified examples are identified and have their emphasis increased

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

what is the key difference between bagging and boosting?

A

at each round, bagging has a uniform distribution while boosting adapts a non uniform distribution.

bagging is parallel, boosting is sequential.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

what is the main boosting algorithm?

A

adaboost

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

what is the update of the distribution function in adaboost after each round?

A

examples from the erroneous portion make up 50% of the next training

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

give the distribution update scheme for adaboost, i.e. what do we multiply Dj(i) by

A

1 / 2ej if the classification was incorrect

1 / 2(1-ej) if the classification was correct

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Generally, an ensemble method works better if the individual models have…

A

Less correlation among predictions

17
Q

Boosting is a family of algorithms, where

A

Each base model has a dataset drawn from a weighted distributed, proportional to the errors made by the previous model

18
Q

Bagging is an ensemble algorithm, where …. is used to ensure predictors are …

A

Bootstrapping is used to ensure predictors are less correlated in their errors

19
Q

where does the name bagging come from

A

bootstrap aggregating

20
Q

list common ML models in order or stability (most to least)

A
svm
knn
logistic regression/perceptrons
neural networks 
decision trees
21
Q

what kinds of models are completely unaffected by bootstrapping

A

stable