Statistical Inference Flashcards

1
Q

Sample mean

A

x¯ = 1/n ∑(i=1, n) xᵢ

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Sample variance

A

s² = 1/(n-1) ∑(i=1 ,n) (xᵢ - x¯)²

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Expected value

A

The expected value of a random variable X is a measure of the centre or average value that one would expect to occur if the random experiment or process were repeated many times. E(X)= ∑ xi ⋅P(X=xi)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Expected value of the sample mean

A

E[X¯] = µ

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Sampling distribution

A

A probability distribution for a sample statistic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Variance

A

Var[x] = E[(x - µ)²] or Var[x] = E[x²] - E[x]²

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Probability density function

A

Probability density function assigns a probability to intervals of values compared to the probability mass function which assigns probabilities to discrete values
PDF is usually denoted f(x)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Cumulative density function

A

CDF is denoted F(x) and is the antiderivative of the PDF f(x)
F(x) = P(X ≤ x)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

PDF of sample maximum

A

for, Z = max(X₁ + X₂ + … + Xₙ)
g(z) = nf(z) (F(z))^n-1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

PDF of sample minimum

A

for, W = min(X₁ + X₂ + … + Xₙ)
h(w) = nf(w) (1 - F(w))^n-1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

CDF of sample maximum

A

G(z) = (F(z))ⁿ or the integral of g(z)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

CDF of sample minimum

A

H(w) = 1 - (1 - F(W)ⁿ or the integral of h(w)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Normal distribution

A

Let X be a normally distributed random variable with mean µ and variance 𝜎².
then we have X ∼ N[ µ, 𝜎²]

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Standardized normal random variable

A

Let X be a normally distributed random variable with mean µ and variance 𝜎².
Then Z = (X - µ)/ 𝜎
and we have mean = 0 and variance = 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is zᵧ, Used to determine the value of x in P(X < x)

A

we denote zᵧ to be the scalar such that P[ Z < zᵧ] = γ. where Z is a random variable that has a standard normal distribution, Z ∼ N [0, 1].

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

when should you use the student’s t-distribution?

A
  • When the sample is small ie. n ≤ 30
    or when the sample standard deviation is given instead of the population standard deviation.
17
Q

Student’s t-distribution

A

If X₁, X₂, …, Xₙ is a random sample from a normal distribution with mean µ and variance 𝜎² then the random variable T = (X¯ - µ)/(S/√n) has the t-distribution with (n-1) degrees of freedom
T∼ t (n-1)

18
Q

what is tᵧ(n-1),

A

We denote tᵧ (n-1) to be the scalar such that P[ T < tᵧ(n-1)] = γ where T ∼ t(n-1)
For example,
n = 11, P[ T < 1.812] = 0.95 t0.95((11)-1) = 1.812

19
Q

Chi-squared distribution

A

The chi-squared distribution is the distribution of the random variable Z²₁ + Z²₂ + . . . + Z²ₙ
If Z is a random variable with chi-squared distribution we can write Z ∼ X²(ν) where ν is the degree of freedom
and X² = (n-1)S²/(𝜎²)
µ = v and 𝜎² = 2v

20
Q

How can we tell its a chi-squared distribution?

A

When the random variable X follows the sum of the squares of independent standard normal random variables.
X = Z²₁ + Z²₂ + . . . + Z²ₙ

21
Q

How can we tell its an F-distribution?

A

The f-distribution is characterised by 2 different sets of degrees of freedom v₁ and v₂.
Use f-distribution when we have two independent random variables X and Y such that X∼ X²(ν₁) and Y ∼ X²(ν₂)
we have F = (X/v₁)/(Y/v₂) of f-distribution F ∼ F(ν₁, ν₂)

22
Q

What is the z score of a poppulation

A

Z = (X -X¯)/(√ (𝜎²/n))

23
Q

Central limit theorem

A

Let X be an (iid) random variable with mean µ and variance 𝜎²
then we have
(a) ∑(i=1, n) Xᵢ ∼ N [nµ , n𝜎²], approximately
(b) X¯ ∼ N[µ , 𝜎²/n], approximately

24
Q

what is an estimator

A

an estimator denoted θ̂ is a statistic used to estimate θ.

25
Q

What is an unbiased estimator

A

an estimator θ̂ of θ is said to be unbiased if E[ θ̂ ] = θ. otherwise it is biased with bias[ θ̂ ] = E[ θ̂ ] - θ.

26
Q

What is the mean square error of an estimator.

A

The mean square error (MSE) of an estimator θ̂ of θ is MSE[θ̂] = E[ (θ̂ - θ)²]
or
MSE[θ̂ ] = Var[ θ̂ ] + {bias[ θ̂ ]}²

27
Q

Better estimator

A

Let θ̂₁ and θ̂₂ be two estimators of a parameter θ,
θ̂₁ is said to be a better estimator in MSE than θ̂₂ if MSE[ θ̂₁ ] < MSE [ θ̂₂ ].

28
Q

Determine the z score of a confidence interval

A

For a 100(1 - 𝛼 ) % confidence level we take z(1-(𝛼/2))
ie. for a 95% confidence level 𝛼 = 0.05.
z(1-(0.05/2)) = Z0.975
then from the z distribution tables find the area of the normal distribution graph to the left of our z𝛼 value.
so z(1-(0.05/2)) = 1.645.

29
Q

Confidence intervals for means, with normally distributed population and known variance

A

P[¯X - Z(1-(𝛼/2)) (𝜎/√ n) < µ < ¯X + Z(1-(𝛼/2)) (𝜎/√ n)] = 1 - 𝛼
where 𝛼 is the amount of risk
or
C = ¯X ± Z(1-(𝛼/2)) (𝜎/√ n)

30
Q

Confidence intervals for means, with normally distributed population and unknown variance

A

P[¯X - t((𝛼/2),n-1) (s/√ n) < µ < ¯X + t((𝛼/2),n-1) (s/√ n)] = 1 - 𝛼
(where s = sample standard deviation.
or
C = ¯X ± t((𝛼/2),n-1) (s/√ n)

31
Q

Poisson distribution

A

Probability mass function for possion distribution is given by P( X = x ) = (𝜆^(x) e^(-𝜆))/x!
The mean and variance for the poisson distribution is 𝜆.
E[ X ] = 𝜆 Var[ X ] = 𝜆

32
Q

Confidence intervals for variance

A

P[S² - S²/(X²(1-(𝛼/2)) < 𝜎² < S² + S²/(X²(𝛼/2)] = 1 - 𝛼

33
Q
A