MODULE 2 S3.2.2 Flashcards
Random Forest (Supplementary)
Decision trees can grow many branches until each split is as __________ as possible.
pure
____________________ are a popular model that mitigates this problem of overfitting in decision trees.
Random Forests
Alternatives to Using a Decision Tree
Random forest
Gradient boosting machine
Support vector machine
Neural network
It is a popular machine learning algorithm that merges the outputs of numerous decision trees to produce a single outcome.
It is also suitable for both classification and regression tasks.
Random Forest
Random Forest’s strength lies in its ability to handle ________________ and mitigate ______________
complex dataset
overfitting
Random forest technique
Bagging
It is a method where multiple machine learning models are trained to solve the same problem and then combined to improve the final output.
Ensemble learning
It is a collection of models is used to make predictions rather than an individual model.
Ensemble learning
Two types of ensemble methods/techniques
Bagging
Boosting
Ensemble Process
Bagging : _______________
Boosting : _______________
Parallel
Sequential
Bagging is known as _____________
Bootstrap Aggregation
It serves as the ensemble technique in the Random Forest algorithm. It is a method of generating a new dataset with a replacement from an existing dataset.
Bagging
How does bagging work?
- Sampling
- Independent training
- Aggregation
Prediction (Bagging Aggregation)
Classification : ______________
Regression : _______________
majority voting
average
Strengths of Bagging
Reduces Variance
Parallelism