Ensemble Learning: Boosting Flashcards

1
Q

What are the key differences between Random Forest and Boosting in terms of tree size and voting?

A

Random Forest: Uses random trees of typical size, where each tree usually has an equal vote in the aggregated decision.
Boosting: Uses stumps (one root, two leaves) where different stumps have different significance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How do Random Forest and Boosting differ in terms of tree independence and order?

A

Each tree is made independently. Order is important, the error of one stump influences the next stump

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the first rule of AdaBoost?

A

Combine stumps (i.e., weak learners)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the second rule of AdaBoost?

A

Some stumps influence the classification more than others.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the third rule of AdaBoost?

A

Each stump is built based on the error in the previous stumps

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the error rate of boosting if the base classifier error rate is less than 50%?

A

If the base classifier (stump) produces a classifier with an error rate < 50%, boosting can give a ~0% error rate classifier.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Describe the basic framework of boosting.

A

1.Obtain the first classifier f1.
2.Find the other function f2 to help f1.
3.Ensure f2 is complementary to f1.
4.Obtain the second classifier f2.
5.Continue this process, and finally, combine all classifiers/

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How is re-weighting of training data done in boosting?

A

-The weight of correctly classified samples is updated as 𝑒(t+1) = 𝑒(t) Γ— 𝑑 or as 𝑒(t+1) = 𝑒(t) Γ— 𝑒^βˆ’Ξ±.
-The weight of incorrectly classified samples is updated as 𝑒(t+1) = 𝑒(t) / 𝑑 or as 𝑒(t+1) = 𝑒(t) Γ— 𝑒^Ξ±.
-Where d = (1-error)/error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the general formula for re-weighting training data?

A

The general formula is 𝑒(t+1) = 𝑒(t) Γ— 𝑒^βˆ’Ε·π‘“(π‘₯)Ξ±, where Ε· and f(x) are Β±1 for binary classification.
-Ε· = f(x): Weight is updated by 𝑒^βˆ’Ξ±.
-Ε· β‰  f(x): Weight is updated by 𝑒^Ξ±

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How are the classifiers aggregated in boosting?

A

Uniform weight: Simply add classifiers βˆ‘f.
Non-uniform weight: Use weighted classifiers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

When is Boosting preferred over Bagging, and vice versa?

A

Useful when subtle features matter (e.g., Face Recognition) and high precision is needed.
*Bagging: Useful when subtle reviews don’t matter (e.g., Sentiment Analysis from Customer Reviews) and there is high variance due to diversity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How does model complexity differ between Bagging and Boosting?

A

Bagging uses complex models like deep decision trees, whereas boosting uses simple models such as stumps

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How does model independence differ between Bagging and Boosting?

A

Bagging trains models independently, whereas boosting builds models sequentially.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How does model aggregation differ between Bagging and Boosting?

A

Bagging uses equal weighting or averaging, while boosting uses non-uniform weights depending on error.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What does each focus on?

A

Bagging is used to reduce variance, while boosting focuses on reducing bias.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly