Week 9: Ensemble Methods Flashcards
Ensemble Methods
A class of methods where a collection of weak classifiers is generated for the dataset, with each weak classifier being a simple classifier. The classifiers are combined together to form the final classifier, by averaging (can be weighted) or voting.
The structures of ensemble classifiers can be parallel, serial, and hierarchical.
Bagging
It’s short for Bootstrap Aggregating. This method uses multiple versions of the training set. With each draw, a different subset of the training set is drawn to train a different weak classifier. The final classification decision is based on votes from each weak classifier.
Boosting
The method begins by making a classifier with a greater-than-average accuracy, and then adding new component classifiers to form an ensemble whose joint decision rule as arbitrarily high accuracy on the training set.
Adaptive Boost (AdaBoost)
This boosting method adds weak classifiers to the final classifier until a desirable classification accuracy is reached. Each added weak classifier is assigned a weight.
Pros:
- Simple to implement
- Good classification accuracy and generalisation
- Can be applied to different classifiers
Cons
- Solution is suboptimal
- Sensitive to noisy data and outliers
Bootstrapping
When data is continuously sampled with replacement to simulate the model viewing unseen data.
Stacked Generalisation
In this method, overall classification is done in 2 levels. The input is processed by group classifiers in the 1st level. Output from the 1st layer will go through the 2nd level classifier to make a final decision.
Random Forest
This method uses an ensemble of decision trees. The final decision is made by combining decisions from all decision trees. Bagging is used on the training data for each tree.