MODULE 2 S3.2.2 Flashcards

Random Forest (Supplementary)

1
Q

Decision trees can grow many branches until each split is as __________ as possible.

A

pure

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

____________________ are a popular model that mitigates this problem of overfitting in decision trees.

A

Random Forests

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Alternatives to Using a Decision Tree

A

Random forest
Gradient boosting machine
Support vector machine
Neural network

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

It is a popular machine learning algorithm that merges the outputs of numerous decision trees to produce a single outcome.

It is also suitable for both classification and regression tasks.

A

Random Forest

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Random Forest’s strength lies in its ability to handle ________________ and mitigate ______________

A

complex dataset
overfitting

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Random forest technique

A

Bagging

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

It is a method where multiple machine learning models are trained to solve the same problem and then combined to improve the final output.

A

Ensemble learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

It is a collection of models is used to make predictions rather than an individual model.

A

Ensemble learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Two types of ensemble methods/techniques

A

Bagging
Boosting

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Ensemble Process

Bagging : _______________
Boosting : _______________

A

Parallel
Sequential

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Bagging is known as _____________

A

Bootstrap Aggregation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

It serves as the ensemble technique in the Random Forest algorithm. It is a method of generating a new dataset with a replacement from an existing dataset.

A

Bagging

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How does bagging work?

A
  1. Sampling
  2. Independent training
  3. Aggregation
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Prediction (Bagging Aggregation)

Classification : ______________
Regression : _______________

A

majority voting
average

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Strengths of Bagging

A

Reduces Variance
Parallelism

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Another ensemble technique, but unlike Bagging, it focuses on reducing bias by sequentially training models.

A

Boosting

17
Q

T/F In boosting, the models created are independent.

A

False

18
Q

T/F In boosting, the models are not independent; they are built one after another, with each model trying to fix the mistakes of the previous ones.

A

True

19
Q

It refers to the family of an algorithm that converts weak learners (base learners) to strong learners.

A

Boosting

20
Q

classifiers that are correct only up to a small extent with the actual classification

A

weak learner

21
Q

classifiers that are well correlated with the actual classification

A

strong learner

22
Q

How does boosting work?

A
  1. Sequential training
  2. Weight adjustment
  3. Final model
23
Q

Strengths of Boosting

A

Reduces Bias
Higher Accuracy

24
Q

Two types of hyperparameters of random forest

A

Increase the Predictive Power
Increase the Speed

25
Q

Predictive power parameters

A

n_estimators
max_features
mini_sample_leaf
criterion
max_leaf nodes

26
Q

Speed parameters

A

n_jobs
random_state
oob_score