Probability Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

Expected value E(x)

A

Also known as the mean
Equal to the sum of all x*f(x)

Example: the expected value of the square of a fair dice roll is 1/6 + 4/6 + 9/6 + 16/6 + 25/6 + 36/6 = 91/6

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Variance V(X)

A

The measure of dispersion or variability.
V(X) = E[(X-μ)^2] = (sum of all f(x)*x^2) - μ^2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Formula linking variance and expected value

A

V(X) = E(X^2) - E(X)^2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Expected value of multiple variables

A

E(X + Y + Z) = E(X) + E(Y) + E(Z)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Variance of multiple variables

A

V(X + Y + Z) = V(X) + V(Y) + V(Z)
independent variables only

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

E(h(X))

A

sum of all h(x)f(x)
So E(x^2) = sum of all x^2
f(x)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

E(aX + b)

A

a*E(X) + b

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

V(aX + b)

A

V(X)*a^2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

P(A or B)

A

P(A) + P(B) - P(A and B)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

P(A given B)

A

P(A and B)/P(B)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Number of ways of choosing r elements from a set of n, and taking order into account

A

P(n, r) = n!/(n-r)!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Our set has n objects, with n1 of one type and n2 of a second type, …, nk of a kth type.
Number of possible permutations?

A

n!/(n1! * n2! * … * nk!)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Number of combinations (unordered arrangements) of r elements in a set of n

A

C(n, r) = n!/(r! * (n-r)!)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Two events are independent…

A

if and only if P(A and B) = P(A)P(B)
==> P(A given B) = P(B given A) = P(A)
P(B)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Bayes’ theorem

A

P(A given B) = P(B given A) * P(A)/P(B)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

General version of Bayes’ theorem for mutually exclusive events

A

P(Ek|A) = P(Ek)P(A|Ek) / [P(E1)P(A|E1) + … + P(En)*P(A|En)]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Bernoulli distribution

A

This is when there’s one random experiment with two possible outcomes
Notation: X~Ber(p), where p=P(X=1)
E(X) = p
V(X) = p(1-p)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Discrete binomial distribution

A

X is the number of successful trials out of n, each with probability p
Notation: X~Bin(n, p)
P(X=k) = C(n, k) * p^k * (1-p)^(n-k)
E(x) = np
V(X) = np(1-p)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Negative binomial distribution

A

X is the number of trials until r successes occur
Notation: X~NegBin(r, p)
P(X=k) = C(k-1, r-1) * p^r * (1-p)^(k-r)
E(X) = r/p
V(X) = r*(1-p)/p^2

20
Q

Discrete geometric distribution

A

X is the number of Bernoulli trials until one success
Notation: X~Geom(p)
P(X=k) = p(1-p)^(k-1)
E(X) = 1/p
V(X) = (1-p)/p^2

21
Q

Standard deviation definition

A

Measure of the amount of variation of a random variable expected about its mean.
Equal to the square root of variance

22
Q

Poisson distribution

A

X is the number of times an event is likely to happen within a specific period of time. λ is the number of times the event is usually expected to happen during that period of time.
Notation: X~Poisson(λ)
P(X=k) = (e^-λ * λ^k)/k!
E(X) = V(X) = λ

23
Q

Using the Poisson distribution to model the binomial distribution

A

𝜆 = np
when n > 20
and np < 5 or n(1-p) < 5

24
Q

Mean of a continuous random variable

A

E(X) = integral from -inf to inf of x*f(x)

25
Q

Variance of a continuous random variable

A

s^2 = V(X) = E[(X-m)^2]
int from -inf to inf of f(x)*(x-m)^2

26
Q

Exponential distribution

A

It is used to model the time elapsing between events, and the probability of occurrence of an event after a period of time. L is the average number of events occurring in a time unit.
Notation: X~Exp(L)
P(X <= k) = 1 - e^-Lx (0 if x < 0)
E(X) = 1/L
V(X) = 1/L^2

26
Q

Discrete geometric distribution

A

All samples in the range have the same probability, meaning
E(X) = (b+a)/2
V(X) = (b-a)^2/12

27
Q

Normal distribution

A

Standard bell curve
Notation: X~N(μ, σ^2)
E(X) = μ
V(X) = σ^2

The area between μ-σ and μ+σ is about 68%, the area between μ-2σ and μ+2σ is about 95%, and the area between μ-3σ and μ+3σ is about 99.7%.

28
Q

Standard normal distribution

A

Special case of the normal distribution with μ = 0 and σ^2 = 1. P(Z <= z) is written as Φ(z)

We convert a normal distribution X to standard ND by setting Z = (X-μ)/σ. We call this the Z-score, a measure of how many standard deviations X is from the mean.

29
Q

Using normal distribution to approximate the binomial distribution

A

When n is large and p is around 0.5.

X~Bin(n, p) => P(X <= x) = Φ((x-np)/sqrt(npq))

30
Q

Using normal distribution to approximate the Poisson distribution

A

X~Poisson(L) => P(X <= x) = Φ((x-L)/sqrt(L))
Good for L > 5

31
Q

Joint probability

A

P(X=x, Y=y) = P(X=x and Y=y) = joint probability of X=x and Y=y

f(x, y) = P(X=x, Y=y)
F(x, y) = P(X <= x, Y <= y)

The JPMF has the following properties: always positive, sums to 1

32
Q

Marginal probability distribution

A

Removing one of the variables from the JPMF
P(X=k) = sum of all the probability pairs (X, Y) where X=k

P(Y=y | X=x) = P(X=x, Y=y)/P(X=x)

33
Q

JPDF of the continuous random variable fXY(x, y)

A

P(a <= X <= b, c <= Y <= d) =
int a to b (int c to d fXY(x, y) dy) dx)

34
Q

Marginal probability distribution: continuous edition

A

fX(x) = int fXY(x, y)dy
fY(y) = int fXY(x, y)dx

35
Q

When are X and Y independent?

A

One of the following statements holds:
fX(x)fY(y) = fXY(x,y)
f(Y=y | X=X) = fY(y) when fX(x)>0
or the other way round
P(X in A given Y in B) = P(X in A)
P(Y in B) for any two sets A and B

36
Q

What is covariance?

A

A measure of the strength of the linear relationship between two variables
cov(X, Y) = σXY = E[XY] - E[X]E[Y]
cov(X, Y) = E[(X-E(X))(Y-E(Y))]

37
Q

Covariance rules

A

Cov(aX+b, cY+d) = acCov(X, Y)
Cov(X1+X2, Y) = Cov(X1, Y) + Cov(X2, Y)
Cov(X, X) = V(X)
If X and Y are independent, then Cov(X, Y) = 0
It’s always between -1 and 1

38
Q

Correlation coefficient

A

Corr(X, Y) = pXY = cov(X, Y)/sqrt(V(X)*V(Y))

39
Q

Markov’s inequality

A

P(X >= a) <= E(X)/a
a > 0

40
Q

Chebyshev’s inequality

A

P(|X- 𝜇| >= k) <= 𝜎^2 / k^2, for all k>0

41
Q

Variance of sample

A

𝜎^2/n
population variance/sample size

42
Q

Central limit theorem

A

Consider n random variables. Suppose they all have distribution F with E(Xi) = 𝜇 and V(Xi) = 𝜎^2, and let X* be the average of all the Xn’s. As n approaches infinity, we will have X~N(𝜇, 𝜎^2 / n). This means that the distribution of the sample mean will be approximately a normal distribution.

43
Q

Bias and variance of an estimator

A

b(t) = E(t) – t
V(t) = E[(t - E(t*))^2]

44
Q

Confidence intervals

A

Depending on the confidence value we want to have, the value of z changes

P(x* - |z|𝜎/sqrt(n) < x < x* + |z|𝜎/sqrt(n)) = 1-a

z = the Z-score when the area to the left is 1 - a/2
(a is small)

Default is two-sided, if one-sided then t <= t* + |z|𝜎/sqrt(n)

45
Q

Hypothesis testing steps

A

Calculate the probability that the sample statistic of a random sample is <= t* (can also be >=), assuming that H0 is true. Call this the p-value.
If the p-value is smaller than a significance level, then reject H0 and accept H1.

46
Q

Test statistic

A

Value used in making a decision about the null hypothesis