stats t1 Flashcards

1
Q

What is a PMF

A

the probability a random variable X = x
it is discrete and must sum to 1.
For continuous data a pdf is used

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is a PDF

A

the probability x falls between the intervals a ands b is the integral F(x) within the limits a and b

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the cdf for discrete data

A

The sum of the PMF is taken for all values of K less than or equal to X

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

what is the cdf. for continuous data

A

P(X<=x) =. the integral of the pdf, from - infinity to x

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what is the expectation of X for discrete data

A

the sum of xP(X), from x

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is the expectation of X for continuous

A

the Integral of x*pmf,
from -infinity to infinity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Var(X) = E[{X − E(X)}^2]

A

= E(X^2) − E(X)^2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the Rth moment

A

E(X^R)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what it the R’th central moment

A

E[(X- mu)^R] where mu =. E(X)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the Co-variance of X and Y

A

E(XY ) − E(X)E(Y )

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is correlation of X and Y

A

Cov(X,Y)/root(var(x)var(y))

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

what are the. linear combination formulas

A

E(aX+bY+c)=aE(X)+bE(Y)+c
Var(aX + bY + c) = a^2Var(X) + b^2Var(Y ) + 2abCov(X, Y )

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

X and Y are independent if

A

fX,Y (x, y) = fX (x)fY (y) or
P(X ≤x,Y ≤y)=P(X ≤x)P(Y ≤y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

for all x and y. If X and Y are independent. then

A

E(XY) = E(X)E(Y) such that
COV(XY) = Corr(XY) = 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

what is the marginal distribution of X

A

The marginal distribution of a random variable X is its probability distribution when ignoring any other variables it may be paired with

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the law of. iterated expectations

A

E(Y) = E[E(Y |X)]

17
Q

uniform distribution

A

P(X=x) = 1/|A| where a is a set
Suitable when all values in A are equally likely.

18
Q

Bernoulli distribution

A

X ∼ Ber(θ), where 0 < θ < 1.
P (X = x) = θ^x(1 − θ)^(1−x)
for x = 0, 1
Suitable when there are only two possible values. Note
that θ = P (X = 1).

19
Q

Binomial distribution

A

X ∼ Bin(m, θ), where m is a positive inte- ger and 0 < θ < 1.

(MCX)* θ^X(1-θ)^M-X
Suitable when X represents the number of ‘suc- cesses’ in m independent Bernoulli trials and θ is the probability of a success on each trial (or analogous situations)

20
Q

Geometric distribution

A

X ∼ Geo(θ), where 0 < θ < 1
P (X = x) = θ(1 − θ)^x
Suitable when X represents the number of failures before the first success in a sequence of indepen- dent Bernoulli trials and θ is the probability of a success on each trial

21
Q

Poisson distribution

A

X ∼ Poi(λ), where λ > 0.

P(X=x)= ((λ^x)e^(-λ))/x!
Suitable when X represents the number of events in a fixed time period, where events occur in- dependently and at random with mean rate λ per unit of time

22
Q

uniform distribution

A

Uniform. X∼Uni(a,b),where −∞<a<b<∞.

f(x) = 1/b-a
for a<x<b.
Suitable when all values in (a,b) are equally likely, or, more formally, all sub-intervals of a given length are equally likely

23
Q

exponential

A

Exponential. X ∼ Exp(λ), where λ > 0.

λe^-λx
Suitable when X represents the time until the first event, where events occur independently and at random with mean rate λ per unit of time

24
Q

normal

A

X ∼ N(μ,σ2), where −∞ < μ < ∞ and σ>0.

f(x) = (1/root(2πσ)) * e^-(x−μ)^2(2σ^2)

A symmetric, unimodal distribution on the whole real line. Suitable when X represents the sum or mean of a large number of random values