Ensemble Learning: Bagging Flashcards

1
Q

What is Ensemble Learning?

A

A machine learning technique that combines multiple individual models to make a final prediction. The idea is that by aggregating the predictions of multiple models, the overall performance of the system can be improved.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the framework behind Ensemble Learning?

A

Obtaining multiple classifiers and aggregating them properly.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are estimators?

A

An estimator (f*) is an approximation of the true function. The goal is to find the estimator from training data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are Decision Trees?

A

Decision Tress are used as base classifiers in bagging and random forests. A decision tree is built by splitting data based on features. Branching criteria determines the splitting of data. The ideal classifier has pure classified sets or least impurity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are two Impurity Measures?

A

Gini Impurity Index and Entropy.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is Gini Impurity Index used for?

A

It is used to measure the impurity of a node. A node is pure (Gini=0) if all instances belong to the same class. The Gini Score is computed as G = 1 - sum of the squares of the ratios of each class.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is Entropy used for?

A

It measures the number of questions you need to ask to get to the data. It is also a measure of uncertainty and information. Higher entropy implies greater uncertainty.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are the advantages of Decision Trees?

A

Each to achieve 0% error rate on training data if each example has its own leaf. Less effort for data preparation (no normalization or scaling needed). Good for model interpretability, making them WhiteBox models.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What are the disadvantages of Decision Trees?

A

High training time and can be expensive. High variance and a tendency to overfit. Instability due to sensitivity to variations in the dataset.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What do Regularization methods do for Decision Trees?

A

They restrict their freedom to prevent overfitting. Examples include: max tree depth, min samples a node must have before a split and min samples a leaf node must have.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is Bootstrapping?

A

A resampling method where random samples are drawn with replacement from a dataset. A random sample of data is selected with replacement from the training set, meaning that data points can be chosen more than once. This random subset is called a bootstrap sample. Multiple models are then trained independently on each bootstrap sample.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is Random Forest?

A

Trains a group of decision tress. Combines results through majority vote, averaging etc. Introduces randomness by using random training sets and feature randomness. Trees are diverse, which decreases training time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is Bagging? (Bootstrap Aggregating)

A

Obtains a set of classifiers. Aggregates the classifiers by sampling N’ examples with replacement from N training examples, usually with N=N’. Separate models are trained on each of these training sets.
Predictions are aggregated through averaging for regression and voting for classification.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are some Bagging advantages?

A

-Reduces overfitting (good for high variance).
-Handles missing values.
-Allows for parallel processing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What are some Bagging disadvantages?

A

-Doesn’t address bias.
-Low interpretability.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Why is bagging effective at reducing variance?

A

The independent models trained on different bootstrapped datasets capture different aspects of the underlying data. By averaging (or voting) their predictions, the model reduces the impact of any single model’s overfitting. This is because errors made by individual models tend to cancel each other out when aggregated, which leads to a more stable and less variable final prediction.

17
Q

What is Out-of-Bag evaluation?

A

In bagging, each model is trained on a different subset of the training data. The data points not included in the bootstrapped sample for training a particular model are called the ‘out-of-bag’ instances. The model’s performance can be evaluated using these ‘out-of-bag’ instances, which provides an estimate of how well the model is generalizing. This eliminates the need for a separate validation dataset.

18
Q

What is the random subspace method?

A

This method further enhances diversity in bagging by randomly selecting a subset of features for each model. This is often used in conjunction with bootstrapping and is a core component of random forests, which also uses a random subset of features at each split point of each decision tree.