7 Flashcards

1
Q

If you have trained 5 different models with 95% precision - how can you improve these

A

Combining them with voting ensemble

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Whats the difference between hard and soft voting classifiers

A

Hard counts votes and picks the most common

soft computes average probability and picks the highest

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Can you speed up bagging, pasting, boosting, random forests, or stacking by distributing across multiple servers

A

Bagging, pasting, and random forests, stacking can be

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Whats the benefit of out of bag evaluation

A

It doesn’t require a validation set for accurate testing - meaning more training data can be used

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Why is extra trees more random, how does it help and is it faster then random forests

A

They use random thresholds as well as random feature subsets

The added randomness acts as a regularization tool and allows for faster training

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How can you make adaboost stop underfitting

A

Increase the number of estimators or reduce regularization hyper parameters or increase learning rate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How can you stop Gradient boosting from over fitting - what can be done with learning rate

A

Decrease learning rate

Early stopping

How well did you know this?
1
Not at all
2
3
4
5
Perfectly