Ensemble Learning Flashcards

1
Q

How do we ensure that learners perform differently? Name five methods.

A

1) Randomly divide original dataset into sub-samples (each learner will have a different subset of data to train on).
2) Randomly select a subset of features (each learner will use different features).
3) Use different control parameter values.
4) Use different training algorithms.
5) Heterogeneous ensembles.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

When do we use the following to combine the output from each learner to produce one final prediction?
a) majority voting
b) average
c) median
d) weighted majority vote

A

a) classification problems
b) regression problems with no outliers
c) regression problems with outliers
d) more complicated problems

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the two main approaches to ensemble learning?

A

1) Bagging (or bootstrap aggregating)
2) Boosting

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Describe the concept of bagging.
(Hint: SMART)

A

Subsamples (learners use different randomly selected subsamples from the dataset).

Multiple learners (several models are trained on the same problem).

Agreement policy (final prediction is based on a predefined policy, like voting or averaging).

Regression and classification (applicable to both types of problems).

Train in parallel (models can be trained independently in parallel).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are the three basic steps of bagging?

A

1) Create multiple subsets of data from original training data.

2) Build multiple models (classifiers).

3) Combine outputs of classifiers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are three approaches to ensure diversity with respect to the predictive behavior of each member in the ensemble?

A

1) create subsets smaller than original by randomly selecting from original.

2) if original dataset small, create subsets of the same size, and sample with replacement.

3) have a large number of members in ensemble (results in different subsets for diversity).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is a weak learner?
Also give three examples of weak learners.

A

A weak learner produces a classifier which is only slightly more accurate than random classification.

A stump is the most common weak learner.
- a stump originates from classification trees.
- it is a tree with one node and two leaves.
- doesn’t make great predictions
- only uses one feature as decision node.

Other weak learners are:
- simple neural networks
- other predictive models.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are five advantages of AdaBoost?

A

1) results in non-linear model.
2) has good generalization performance.
3) robust to overfitting.
4) simple to implement
5) works with numerical and nominal features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is a weakness of AdaBoost?

A

Sensitive to outliers (outliers might have an increased weight for next learner).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Once trees in a forest are induced, then the final prediction can be determined by: ____

A

1) majority voting (classification)
2) weighted majority voting
3) average across trees in forest (regression)
4) median (regression with outliers)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are four advantages of random forests?

A

1) performs well on most problems - especially if trees are diverse.
2) can handle missing values and noise.
3) efficient on large number of features.
4) pruning is not needed (no need to worry about overfitting).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What are three weaknesses of random forests?

A

1) not easy to interpret.
2) rule extraction is not possible.
3) some tuning of parameters might be needed.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Why are decision trees popular for ensembles?

A

Sensitivity to dataset changes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How do diverse weak learners in an ensemble boost performance?

A

By compensating for each other’s weaknesses.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

True or False: Bagging is simpler and easier to parallelize than boosting.

A

True.
Bagging is more user-friendly and faster to train.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Boosted ensembles are prone to overfitting. How can this be mitigated?

A

Learning rates.

17
Q

How can ensembles be applied to classification tasks?

A

Through voting via majority class.

18
Q

How can ensembles be applied to regression tasks?

A

Central measures of tendency.