ML learn Flashcards

1
Q

List xgboost benefits

A

L1l2 reg prevents overfitting on high dimensional space
Missing values handling
Cross-validation
Allow early stopping
Option to look at the learning graph to choose different checkpoint
Multiple cpu optimised
Goes to deep trees -> allows more optimised trees for inference
Allow multiple objective functions
Easy interface (python)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

XGBoost init

A

Data and a differentiable loss function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

XGBoost step 1

A

Init the model with a constant value : F0=argmin (gamma) SUM i=1..n (yi, gamma)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

XGBoost step 2 headline

A

m=1 to M

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

XGBoost step 2 a

A

compute r_im = [diff(L(yi, F(xi))/diff(F(xi))] when F(xi)=F_m-1(xi) for i=1..n

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

XGBoost 2 b

A

Fit a regression tree to the $r_{im}$ values and create terminal regions $R_{jm}$, for $j = 1, \dots, J_m$

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

XGBoost 2 c

A

For $j = 1, \dots, J_m$, compute
[
\gamma_{jm} = \arg\min_\gamma \sum_{x_i \in R_{jm}} L(y_i, F_{m-1}(x_i) + \gamma)
]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

XGBoost 2 d

A

Update
[
F_m(x) = F_{m-1}(x) + \nu \sum_{j=1}^{J_m} \gamma_{jm} \mathbb{I}(x \in R_{jm})
]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

XGBoost 3

A

Output Fm(x)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

XGBoost start - explain in words what is the initial prediction

A

The value that minimises the loss function over the observations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

XGBoost - explain in words how to refine the prediction over the previous prediciton

A

Add a tree that minimises the errors of the previous predictions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

XGBoost - explain in words how to add the new tree prediction that minimises the errors of the previous prediction

A

Scale the tree with the learning rate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q
A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly