Probability and statistics Flashcards
Unbiased Estimstor
If E(Tn) = θ
Tn is an unbiased estimator
Bias
b(Tn, θ) = E(Tn) - θ
The bias of Tn as an estimator of θ
Sampling Distribution
The distribution of the estimator
Sampling Error
The standard deviation of the sampling distribution
The sd is the sampling sd = S
Consistent Estimator
If Tn is unbiased for θ, Var(Tn) → 0 as n → ∞
∀∈>0 P( |Tn - θ|
Mean Square Error
MSE(T) = Var(T) + bias squared
MSE(T) = E[(T-θ)²]
Proof: MSE = Var(T) + b²(T,θ)
E[(T-θ)²] = E[(T-E(T) +E(T)-θ)²]
= E[(T-E(T))² + 2(T-E(T)) (E(T)-θ)+(E(T-θ)²]
=Var(T) + 2(E(T)-θ) E(T-E(T)) + (E(T) - θ)² = Var(T) + b²(T,θ)
Coefficient of Variation
S.d. / E(x)
Null Hypothesis
An assumption about a parameter which we wish to test in the basis of available data - H₀
Alternative Hypothesis
If the data are not deemed to support H₀ then we will conclude than an alternative hypothesis H₁ is supported
P-Value
Observed significance level
The observed significance level of a test (p-Value)
The probability of obtaining a value of the test statistic at least as extreme as that observed under H₀
Type 1 Error
If H₀ is true and we reject it
Type 2 Error
If H₀ is false and we accept it
Power
The quantity 1-β is called the power of a statistic test
It measures the test’s ability to detect a departure from H₀ when it exists
Critical value
The value below which we accept H₀ and above which we reject it
Komogorov’s Axioms of Probabilty 1-3
A probability function Pi is a mapping P:F→ℝ s.t.
- ∀ E∈ F, P(E) ≥ 0
- P(Ω) = 1
- If E ∩F = ∅ then P(E)∪ P(F) = P(E) + P(F)
Deductions from the axioms 1-4
- P(E°) = 1 - P(E)
- P(E) ≤ 1
- If E ⊆ F, the P(F|E) = P(F) - P(E)
- For any events E and F, (not necessarily disjoint); P(E∪F) = P(E) + P(F) - P(E∩F)
Independence
Events E and F are independent if P(E∩F) = P(E)P(F)
E and F are unrelated - E doesn’t affect F
Mutually Independent
Pairwise independence
Events E1…En are mutually independent of for any collection of the events, the independence relation holds
All pairs of events Ei and Ej are independent
Conditional Probability
The conditional probability of F given E is the probability of F occurring when E is known to have occurred
Sample space changes from Ω to E
For independent events - P(F) given E is just P(F)
Law of Total probability
Let {Ei} be a partition of Ω st. Each outcome belongs in exactly one of the partitions
Then P(F) = ∑P(F|Ei)P(Ei)
Proof:
F = F∩Ω = F∩(∪iEi) = ∪i(F∩Ei) = P(F)
=∑P(F∩Ei) = ∑P(F|Ei)P(Ei)
Random variable
A function from a sample space Ω to ℝ
Discrete: a function from a countable sample Ω to ℝ: X:Ω → ℝ is a d.r.v.
Continuous: X is a c.r.v. If Fx is continuous and differentiable
Probability Mass Function - drv
If x is a d.r.v. Taking values in the set {xi}, then the function Px(x) = P(X=x) is the pmf of x
Properties:
1. P(Xi) ≥ 0 for all i since these are probabilities of events (axiom 1)
- ∑P(Xi) = 1
Axiom 3
Probability Density Function - CRV
He pdf is fx - the derivative of the distribution function Fx
Properties:
1. f(x) ≥ 0
- The integral of f(x) over R is 1
Cumulative distribution Function
For any random variable X, the function Fx(X) = P(X≤x) = ∑ Px(xi)
Drv: F(X) is a step function with discontinuities at the Xi
CRV: Fx is continuous and differentiable
Properties:
- F(x) → 1 as → x
- F(x) → 0 as → x
- F(x) is monotonic increasing x1
Expectation Function
Properties:
E(x) exists if:
- sample space is finite
- the sum / integral converges absolutely
- Idealised long run average
Discrete:
Sum ∑xi P(Xi)
Continuous: Integral xf(x) over R
E(x) - sum / integral of xi times pmf / pdf
Properties of E(x)
- X = constant - P(X=c) = 1, then E(X) = C
- Y = aX +b
E(Y) = aE(X) + b
Proof - summation / integral and compute
Symmetry of E(x):
X has a symmetric pmf/pdf - if E(x) exists it is the central point of the pmf/pdf
If symmetric about μ, Let Y = X-μ - pmf is not symmetric about 0 so E(Y) = 0, E(X) - μ = 0 rearranging gives result
Properties of variance
- Var(x) ≥ 0
Sum of positive terms - all squared and real
- Var(x) = 0 if X is constant is. p(x=c) =1
Compute
- y = aX + b, var(y) = a²var(x)
Compute
Coefficient of Variation
δ/μ
The ratio of s.d / mean
Bernoulli
An experiment with 2 outcomes: success and failure
X - only takes values 0 or 1
Binomial
Sum of n independent Bernoulli trials
X - no. Of successes in n independent Bernoulli trials with probability of success p
Geometric
X - no. of independent Bernoulli trials until a success
The waiting time between successes in binomial
Negative Binomial
X - no. of Bernoulli trials until rth success
The sum of r independent geometric