Lecture 8 Flashcards

1
Q

What level of variance and bias is the most fundamental (high/low)?

A

Low bias and low variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Which models do not perform well well by themselves due to high bias or high variance?

A
Logistic Regression
Naive Bayes
KNN
(Shallow) Decision Trees
Linear SVMs
and Kernel SVMs
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What ways can we combine models?

A

Voting
Bagging
Boosting
and Stacking

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Bagging

A

Trains many models on bootstrapped data, then takes the average

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Boosting

A

Given weak models, it runs it multiple times on reweighted training data, then lets learned classifiers vote with Gradient Boosting

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Stacking

A

Trains many models in parallel and combines them by training a meta-model to output a prediction based on the different weak models predictions.

The training data is split into two folds. One fold is used to train the base models and the other is used to train the meta-model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Random Forest approach

A

a bagging method where deep trees, fitted on bootstrap samples, are combines to produce an output with lower variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What does Bagging aim to reduce?

A

variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does Boosting aim to reduce?

A

bias

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Bootstrapping

A

a resampling technique that generates samples of size N (called bootstrap samples) from an initial dataset by randomly resampling ( drawing with replacement) N observations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly