12. Ensemble Method* Flashcards
decision trees
while easy to implement, train, use and interpret, it is inaccurate
core idea of ensemble method
when weak models are carefully trained and combined, they should be able to produce more accurate and robust prediction results
what are the ensemble methods
- bagging (bootstrap aggregation)
- create diverse sample data with replacement
- train model with subset of data and aggregate -> learn B weak learners
- aggregate predictions from all learners
- achieve good results by reducing variance - boosting
- training a sequence of weak models, each compensating the weaknesses of its predecessors
- correct predicting error
- achieve good results by decreasing the bias - stacking
- train multiple models from the same training dataset and use another model to aggregate prediction
random forest
improve variance deduction of bagging by reducing the correlation between base models without increasing the variance
p * sigma^2 (1 - p) * sigma^2 / n
how does random forest work..?
adaboost**
put more weight on difficult instances and less on those already learned
- set weights of all samples
- bootstrap k samples based on weights
- fit leaner XX on sample data
- evaluate weak learner G on dataset and calculate loss
- compute impacting factor of alpha
- update sample weights based on e
- update G
- repeat till terminate
gradient boost**
learn the errors (pseudo residuals) made from predecessors
what are the advantages to ensemble
- able to learn complex problems
- generally high accuracy
what are the disadvantages to ensemble
- might not fit well in simple problems
- lack of interpretability
- computationally expensive