Chapter 10- Ensemble Models Flashcards
what is the principle of committee decision?
individual predictions combined appropriately should have better overall accuracy on average than an individual committee member
what three methods can be used to combine decisions?
averaging, voting, probablistic
what is the wisdom of crowds?
different versions make errors in different ways. combining them cancels out the errors of the individuals
what is bagging?
generate different training sets, by sampling with replacement to build a committee of models
is bagging a sequential or parallel method?
parallel
what are random forests?
an extension of bagging applied to decision trees.
what are the two methods of generating randomisation for random forests?
bootstrap, as in bagging
random selection of features are each split point
why do we force a random selection of feature at each split point in random forests?
to ensure it doesnt choose the best feature at every split.
what does the lack of pruning of random forest trees ensure
that the models aren’t too simple and therefore too similar to each other.
is adaboost parallel or sequential?
sequential
describe the adaboost algorithm
a new model trained at each round. at the end of each round the mis-classified examples are identified and have their emphasis increased
what is the key difference between bagging and boosting?
at each round, bagging has a uniform distribution while boosting adapts a non uniform distribution.
bagging is parallel, boosting is sequential.
what is the main boosting algorithm?
adaboost
what is the update of the distribution function in adaboost after each round?
examples from the erroneous portion make up 50% of the next training
give the distribution update scheme for adaboost, i.e. what do we multiply Dj(i) by
1 / 2ej if the classification was incorrect
1 / 2(1-ej) if the classification was correct