Chapter 2- Probabilistic Models Flashcards

1
Q

what is a random variable?

A

a numeric quantity whose values map to the possible outcomes of an experiment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what is a sample space, or alphabet?

A

consists of all possible events

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

what is an event?

A

any subset of values that X can take

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

what is the first rule of probability theory?

A

probabilities add up to 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what is the difference between estimated and true probabilities?

A

an estimate comes from a sample- it is a sample estimate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is another name for the true probability?

A

a population parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what is joint probability?

A

AND

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

from conditional probabilities, p(x,y) = ?

A

p(x|y)p(y) and vice versa, p(y|x)p(x)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what is the rule for independent events?

A

p(x,y) = p(x)p(y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

how is bayes theorem derived?

A

p(x,y) = p(x|y)p(y)
and vice versa p(x,y) = p(y|x)p(x)
equate these

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

what is another word for joint probability?

A

marginalisation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

what is the formula for joint probability, p(X=x)?

A

sum for each y: p(x|y)p(y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

what is the conditional independence assumption? (in words)

A

the features are conditionally independent of each other, given the class value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

what is the conditional independence assumption p(x1,x2,x3|y) = ?

A

p(x1|y)p(x2|y)p(x3|y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

what is the conditional independence assumption p(X|y) = ?

A

multiply for each x: p(x|y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

what is the naive bayes model?

A

we solve problems by making the conditional independence assumption

17
Q

a bayesian network is an example of?

A

a directed acyclic graph

18
Q

when might over/underfitting occur in a bayesian network?

A

if we have more links, the more complicated the probability distribution and hence more data is needed

19
Q

what is one of the great advantages of bayesian networks?

A

you can very naturally encode human knowledge about a given problem

20
Q

what is the chain rule for probability: P(A n B) = ?, and P(A1 n A2 n A3 n A4) = ?

A

P(A n B) = p(B|A)P(A)
or for more
P(A1 n A2 n A3 n A4) = p(A4 | A1 n A2 n A3)
= P(A4 | A3 n A2 n A1) p(A3 | A2 n A1)p(A2 | A1) p(A1)

21
Q

what does the chain rule tell us how to compute? Events A and B

A

The chain rules tells us how to compute the probability of event A happening AND then event B occurring afterwards.

22
Q

what is a random experiment?

A

random experiment is any experiment in which the outcome is uncertain (not known or determined in advance).

23
Q

what is P(A or B) if

a) they are disjoint
b) they are joint

A

a) P(A) + P(B)

b) P(A) + P(B) - P(A and B)

24
Q

what is a probability mass function?

A

discrete random variables take on a finite number of values.
Each value is associated with a probability of it occurring. t
The collection of these probabilities is the probability mass function

25
Q

give the bernoulli distribution

A
P(X = 0) = 1 - p
p(X = 1) = p
26
Q

give the binomial distribution, P(X=k) =

A

P(X = k) = (nCk)(p^k)(1-p)^(n-k)

27
Q

when would we use the binominal distribution

A

if experiment is repeated n times, the probability that we will see k successes is given by this probability mass function

28
Q

give the geometric distribution

A

P(X=x) = (1-p)^x-1 (p)

29
Q

when would we use the geometric distribution

A

it is useful for modelling the first occurrence of an outcome after repeated identical trials

30
Q

give the poisson distribution

A

P(X=x) = { lambda^x e(-lambda) } / x!

31
Q

when would we use the poisson distribution

A

if we had a rate e.g. 5 times a year

32
Q

if a discrete r.v. X has a pmf f(X) what is the expected value E[g(x)]

A

sum i: g(Xi)f(Xi)

33
Q

if a discrete r.v. X has a pmf f(X) what is the variance V[g(x)]

A

E[(g(X) - E(g(X)))^2]

E[g(X)^2] - E[g(X)]^2

34
Q

properties of Expectations

E[aX + b] =

A

aE[X] + b

35
Q

properties of variance:

V[aX+b] =

A

a^2V[X]

36
Q

what are the two schools of probability?

A

frequency based

belief based

37
Q

Describe the gamblers fallacy and how it demonstrates iid

A

A sequence of outcomes of spins of a fair or unfair roulette wheel is i.i.d

The outcome of the previous turn doesnt impact the next turn

The distribution of probabilities is the same each time i.e. 50/50 if not biased.

So, even if we’ve had 20 reds, there is still the same chance of having a black vs red next go

38
Q

Let the two events be the probabilities of persons A and B getting home in time for dinner, and the third event is the fact that a snow storm hit the city. If they live in different areas, are A and B conditionally independent given C?

A

Yes. That is, the knowledge that A is late does not tell you whether B will be late.

If they lived in the same neighbourhood they would not be.

39
Q

if two events are independent, are they also conditionally dependent?

A

not necessarily, it depends on the third event.

rolling a dice twice are two independent events. if the third event was the sum of them being 7, they would then be independent but not conditionally independent given c