Topic 6: Ensemble Theory Flashcards

1
Q

Bias-Variance-Diversity decomposition

A

Expected risk (ensemble) = noise + bias + variance - diversity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is Diversity

A

ESn [ 1/m mΣi=1 (fi(x) - f¯(x))^2 ]

Difference between models
diverse models make different errors on new data points

compares prediction made my ith model fi(x) and the average over all models

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How does diversity help in ensembles

A

computational efficiency
robustness against adversarial attacks
improved performance in various applications

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is centroid * (circle above term)

A

Can refer to any of: arithmetic mean, harmonic mean, etc
Represents the centre of the model distribution
It is averaging over all possible data sets (infinite)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Generalised Bias-variance decomposition

A

ED [ EXY [ ℓ(Y,q)] ] = EX [ EY|X [ ℓ(Y,Y)] ] + ℓ(Y*,q◦) + ED[ ℓ(q◦, q)
Expected risk = noise + bias + variance

{z }

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Generalised ambiguity decomposition

A

ℓ(y, q¯) = 1/m Σ ℓ(y, qi) - 1/m Σ ℓ(q¯, qi)
Ensemble loss = average loss - ambiguity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is q¯

A

The ensemble combination
Eg for squared loss ->arithmetic mean
for KL -> normalised geometric mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How does regularisation effect variance and bias

A

It moves models around in the variance bias axes
It can decrease variance and may increase bias
Eg Linear + reg = lower var, higher bias

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Larger v smaller networks and diversity

A

Larger networks tend to perform better due to lower bias and variance, despite potentially lower diversity(capture more complexity and overfit) compared to smaller networks

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How does diversity interact with bagging

A

Random Forests initially underperform Bagging but catch up as the ensemble size increases
Random Forests exhibit higher variance-effect but compensate with higher diversity-effect in larger ensembles

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the centroid combiner for poisson regression loss

A

geometric mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the centroid combiner for KL divergence

A

normalised geometric mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the centroid combiner for itakuro saito loss

A

harmonic mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

what is the centroid combiner for squared loss

A

arithmetic mean

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

single model vs ensemble model tradeoffs

A

In single models we have a 2-way tradeoff (bias/variance)

In ensembles of models, it’s a a 3-way tradeoff (bias/variance/diversity)

But it only holds if we use the centroid combiner rule

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the key of diversity decomposition

A

it is not the task (e.g., classification/regression) that matters, but the loss function
it is a “hidden dimension”

17
Q

What is diversity dependent on

A

The label (due to combiner)

18
Q

How is diversity always related to expected risk

A

diversity always subtracts from the expected risk
But it only reduces it given a fixed bias and variance

19
Q

what is y bar

A

average prediction made by all models in the ensemble for a given input x

20
Q

does bias-variance decompositions hold for all losses

A

No

21
Q

difference between q bar and q circle

A

q bar
represents the simple average
m∑i=1 qi
used in diversity

q circle
represents expected value, takes into account probabilities and weights
used in bias and variance