Theory of Learning from Data Flashcards

1
Q

Was ist eine Risk Function?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Wie unterscheiden sich true (expected) risk und empirical risk?

A
  • true expected risk ist das Risiko/der Verlust auf unbekannten Datan
  • empiricak risk ist das Risiko/der Verlust auf bekannten Daten
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Was bedeutet VC(H)?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Gib VC(H) an: Linear classifiers for d features plus a constant term b

A

d + 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Gib VC(H) an: Decision tree of rank r that defines Boolean functions
on n boolean variables

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Gib VC(H) an: Neural networks

A

VC(H) ≈ #parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Gib VC(H) an: Linear classifier in 2D mit drei Punkten

A

VC(H) = 3

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Was ist Structural Risk Minimization?

A

Risk Calculation of different Models

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Wie geht der Satz von Bayes?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Wie unterscheiden sich bayesian view und cost funtion view?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Gib die Bayesian probabilistic formulation

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Wie hängen bayesian view und cost function view zusammen?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Wie kann man Modellkomplexität verringern?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Wie kann man Parametergrößen “restricten” (beschränken)

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Beschreib Regularizer (L2 norm)

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Beschreib Regularizer (L1 norm)

A
17
Q

Erkläre Cross Validation

A
18
Q

Which of the following statements on the different kinds of cross-validation are correct?
1. The leave-one-out method is a special form of k-fold cross-validation.
2. Cross-validation is used to find the best training data to train a model.
3. The bootstrap resampling technique involves dividing the dataset into multiple partitions, evaluating each subset individually as test data after training on the rest.
4. A major advantage of k-fold cross-validation is that it is a fast method to test the quality of the chosen model.

A

1

19
Q

The Vapnik Chervonenkis (VC) dimension of a classifier H is the cardinality of the smallest set that can be fully represented by H.
Ist das Wahr?

A

Nein, actually it is the largest set a classifier H can fully represent.

20
Q

Which of the following statements on VC theory are correct?
1. A larger model complexity implies a smaller empirical risk.
2. The effective model complexity is fixed during the course of training.
3. The empirical risk is a good measure for the generalization capabilities of a model.
4. Structural risk minimization balances empirical risk and VC dimension.

A

1 und 4