PAC Flashcards

1
Q

PAC?

A

Probably Approximately Correct Learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is Sample Complexity?

A

The class of learnable concepts in terms of the number of sample points needed to achieve an approximate solution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is “X” or Input Space?

A

The set of all possible examples or instances.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is “Y”?

A

The set of all possible labels or target values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is Concept?

A

A concept c is a mapping rom X to Y - X -> Y. For example all the points which fall into a triangle.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is Concept Class?

A

Denoted as C. It is the set of concepts we may wish to learn. E.X be the set of all axis-aligned rectangles.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is a Hypothesis Set?

A

Denoted as H. It is a fixed set of possible concepts which may not coincide with C.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is Generalization Error?

A

Generalization error of hypothesis h also known as true error or just error is denoted by R(h) (risk).
R(h) = Pr [h(x) != c(x)] = E 1_ h(x) != c(x)
The Generalization error is not directly accessible to the learner since both the distribution D and the target concept c are unknown. GR is the expected error based on the distribution D. The expectation of the empirical error R^(h) is also the generalization error = E[R^(h)] = R(h).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is Empirical Error?

A

R^(h) = 1 / m * SUM(i=1…m) 1 if h(xi) != c(xi) else 0

Thus Empirical error is the average error over the sample S.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the definition of PAC-learning?

A
A concept class C is said to be PAC-learnable if there exists an algorithm A and a polynomial function P such that for any epsilon > 0 and delta > 0 for all distributions D on X and for any target concept c WBI C, the following holds for any sample size m >= poly(1/eps, 1/delta, n, size(c))
Pr( [R(h_S) ) <= eps] >= 1 - delta
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the term epsilon e?

A

Its the error term. It tells that we are approximately correct. How accurate the algorithm is (1 - eps).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the term delta d?

A

It tells that we are probably right (1 - delta). In other words the confidence of the algorithm.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What are false negatives?

A

Points that the hypothesis misses that are actually inside the concept class but not inside the hypothesis.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are false positives?

A

Points that the hypothesis flags as being class A but in truth are not. So they are not inside the concept class A but are inside the hypothesis.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is a consistent hypothesis?

A

It is a hypotheses which has empirical error R^(h) = 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the version space?

A

The set of all consistent hypotheses of the hypothesis class.

17
Q

What is the most general hypothesis G?

A

It is the hypothesis which cannot be expanded without including negative training examples. “The largest rectangle”

18
Q

What is the most specific hypothesis S?

A

It is the hypothesis which cannot be made smaller without excluding positive training points. “The smallest rectangle”