Stats Flashcards
What is a UNIFORM Probability Space?
A uniform probability space is defined as the triplet (Ω, F, P),
where
* Ω (the sample space) is a non-empty finite set of all possible outcomes of the experiment;
* F (the event space) is the collection of all events, given by the power-set P(Ω) of Ω;
* P : F → [0, 1] is a map from the event space to [0, 1], satisfying:
− P(∅) = 0, P(Ω) = 1
− P(A ∪ B) = P(A) + P(B), for every A, B ∈ F such that A ∩ B = ∅
(finite additivity)
− P({ω}) = P({ω˜}) for all ω, ω˜ ∈ Ω (uniform)
How do we calculate probability in a uniform probability space?
Let (Ω, F, P) be a uniform probability space. Then for all ω ∈ Ω,
P({ω}) = 1/|Ω|
and for all A ⊆ Ω (A ∈ F),
P(A) = |A|/|Ω|.
Proof,
Since ∀ ω1, ω2 ∈ Ω, P({ω1}) = P({ω2}), let p ∈ [0, 1] s.t.
p := P({ω}) ∀ ω ∈ Ω .
Since P is a probability measure
1 = P(Ω) = Σω∈Ω P({w}) by (1.1b)
= Σω∈Ω p
= pΣω∈Ω 1
= p|Ω|.
Therefore
p = 1/|Ω|.
showing (1.2). We see that (1.3) follows from (1.1b), as
P(A) = Σω∈A P({ω}), by (1.1b)
= Σω∈A p
= pΣω∈A 1
= p|A| = |A|/|Ω|.
What are Sequences of a set?
Let A be a finite set of cardinality n ∈ N. A sequence of length
k ∈ N of elements of A is an ordered k-tuple (a1, . . . , ak) s.t. ai ∈ A, i = 1, 2, . . . k.
We denote by Sn,k(A) the set of all sequences of length k of elements of A.
By “ordered” we mean that the order of the sequence matters, for example:
(a1, a2) /= (a2, a1). Note also that repetitions of elements are allowed, for example
(1, 1) is a sequence of length 2 of elements of {1}.
What is the Cardinality of Sn,k?
Let A be a finite set of cardinality n ∈ N. The set Sn,k(A) of
all sequences of length k ∈ N of elements of A has cardinality n^k, i.e.
|Sn,k(A)| = n^k
Proof,
To construct an arbitrary element (a1, a2, . . . , ak) of Sn,k(A), we perform the following steps
(a) choose the first value a1. There are n1 = |A| = n ways to do this.
(b) choose the second value a2. There are n2 = |A| = n ways to do this.
.
.
.
k) I choose the kth value ak. There are nk = |A| = n ways to do this.
Thus we find
|Sn,k(A)| = n1 · n2 · · · · · nk = n · n · · · · · n = n^k.
What is an Ordering of a set?
Let A be a finite set of cardinality n ∈ N and let k ∈ N such
that k <= n. An ordering of length k of elements of A is a sequence of length k of elements of A with no repetitions. We denote the set of orderings of length k of elements of A by On,k(A). Thus we have
On,k(A) = {(a1, . . . , ak) : ai ∈ A ∀i = 1, . . . , k, ai /= aj ∀ i /= j}.
What is the Cardinality of On,k?
Let A be a finite set of cardinality n ∈ N and k <= n. Then
|On,k(A)| = n(n − 1). . .(n − k + 1).
Proof. We determine an element of On,k(A) by the following steps:
(a) We choose a1. There are n1 = |A| = n choices for this.
(b) We choose a2 such that a2 /= a1. There are n2 = n − 1 choices for this.
(c) We choose a3 such that a3 /= a2 and a3 /= a1. There are n3 = n − 2 choices
for this.
.
.
.
k) We choose ak such that ak /= ai ∀ i = 1, 2, . . . , k − 1. There are nk = n − (k − 1) choices for this.
By the fundamental counting principle
|On,k(A)| = n1n2 . . . nk = n(n − 1). . .(n − k + 1).
What is a Combination of a set?
Let A be a finite set of cardinality n ∈ N. A combination of k
elements of A is a subset of A with k elements. We denote by Cn,k(A) the set of combinations of k elements of A.
What is the Cardinality of Cn,k?
Let A be a finite set of cardinality n ∈ N and k <= n. Then
|Cn,k(A)| = nCk = n!/[k!(n − k)!].
Proof. Notice that an ordering of length k of elements of A can be obtained uniquely by the following steps
(a) Choose a combination of k elements of A. There are n1 = |Cn,k(A)| choices for this.
(b) Choose a permutation of these elements. By Corollary 2.1 there are n2 = k! choices for this.
By the fundamental counting principle,
|On,k(A)| = |Cn,k(A)| · k!.
The above equation can be solved for the only unknown term, giving
|Cn,k(A)| = |On,k(A)|/ k! = n!/(n − k)!k!
= nCk.
What is a Permutation and how many permutations of a set are there?
Let A be a finite set of cardinality n ∈ N. An ordering of length
n of elements of A is called a permutation of A.
Let A be a finite set of cardinality n ∈ N. Then the number of
permutations of the elements of A is n(n − 1)(n − 2). . . 1 = n!.
Proof, just take k=n as an ordering.
What is a Partition of a set?
Let A be a set of cardinality n ∈ N and r ∈ N such that r <= n.
A partition of A into r subsets is a family {A1, . . . , Ar} of subsets of A such that
(a) Every subset in the family is non-empty: Ai 6= ∅ ∀ i = 1, 2, . . . , r.
(b) The subsets in the family are pairwise disjoint: Ai ∩ Aj = ∅ ∀ i /= j, i, j = 1, 2, . . . , r.
(c) The union of all the subsets in the family is equal to A:
Ui=1,r (Ai) = A.
How many partitions of a set are there?
Let A be a finite set with |A| = n. Let r ∈ N, r <= n. Then
the number of partitions of A into r subsets {A1, . . . , Ar} such that |A1| = k1, |A2| = k2 . . . , |Ar| = kr is given by:
n!/k1!k2! . . . kr!.
Proof. Every partition of A satisfying the assumptions can be uniquely determined via the following steps:
(a) Choose A1 ⊆ A such that |A1| = k1. There are nCk1 choices for this step.
(b) Choose A2 ⊆ A such that |A2| = k2 and A1 ∩ A2 = ∅ (which implies that A2 ⊆ A \ A1). There are n−k1Ck2 choices for this step.
(c) Choose A3 ⊆ A such that |A3| = k3, A3 ∩ A2 = ∅ and A3 ∩ A1 = ∅ (which implies that A3 ⊆ A \ (A1 ∪ A2)). There are n−k1−k2Ck3 choices for this step.
.
.
.
r) Finally choose the remaining set Ar ⊆ A such that |Ar| = kr and Ai∩Aj = ∅ for all i = 1, 2, . . . , r − 1 (we see that Ar ⊆ A \ (A1 ∪ · · · ∪ Ar)). There are n−(k1+k2+···+kr−1)Ckr choices for this step.
Thus, by the fundamental counting principle the number of partitions of A into r subsets {A1, A2, . . . , Ar} such that |A1| = k1, . . . , |Ar| = kr with k1 + k2 + · · · + kr = n is
nCk1 * (n − k1)Ck2 * . . . * n − (k1 + · · · + kr−1)Ckr
=n!/k1!(n − k1)! * (n − k1)!/k2!(n − k1 − k2)! * . . . * (n − (k1 + · · · + kr−1))!/kr!(n − (k1 + · · · + kr))!
= n!/k1!k2! . . . kr!,
where the last equality comes by cancellation of the fractions and noticing that
(n − (k1 + · · · + kr))! = (n − n)! = 0! = 1.
What is a Sample Space?
A sample space Ω is the set of all possible outcomes of a random process (or experiment). It can be any set.
What is an Event Space and what are the known Properties of an Event Space?
Let Ω be the sample space and F be a collection of subsets of
Ω. F is an event space (also called σ-algebra) if it satisfies
(i) Ω ∈ F.
(ii) if A ⊆ Ω, A ∈ F ⇒ Ac ∈ F (F is closed under compliments).
(iii) if {An : n ∈ N} is such that An ∈ F ∀ n then
Un=1,∞ (An) ∈ F . (F is closed under countable unions.)
Let F be a event space on Ω. Then
(a) F is closed under finite unions.
(b) F is closed under finite intersections.
(c) F is closed under countable intersections.
Proof.
(a) Let A1, . . . , An ∈ F. Set Aj = ∅ ∈ F ∀ j > n, thus An ∈ F ∀ n > 1. Since the empty sets will not contribute to the union, we can show that
Uj=1,n Aj = Uj=1,∞ Aj ∈ F,
as F is closed under countable unions.
(b) Let A1, . . . , An ∈ F. We want to show that ∩j=1,n Aj ∈ F. By De Morgan’s law (show that every element of one set needs to belong to the other set as well)
∩j=1,n Aj = (Uj=1,n Aj^c)^c.
Since the event space F is closed under taking complements, Aj^c ∈ F, for every j = 1, . . . , n. Since it is closed under finite unions (statement 1 of the proposition, shown above),
Uj=1,n Aj^c ∈ F
By taking complement once more, it follows that ∩j=1,n Aj ∈ F.
(c) The proof is similar to that of statement 2 above, noting that De Morgan’s law also holds for countable unions and intersections. That is, we can write
∩j=1,∞ Aj = (Uj=1,∞ Aj^c)^c.
What is a Probability Measure?
Given a sample space Ω and an event space F, a function P : F → R is called a probability measure if it satisfies
(i) P(B) ∈ [0, 1] for every B ∈ F;
(ii) P(Ω) = 1;
(iii) (Countable additivity) For every An ∈ F, n >= 1 disjoint events
P(Un=1,∞ (An)) =Σn=1,∞ (P(An)).
What is a Probability Space?
A probability space is defined as the triplet (Ω, F, P), where
* Ω (the sample space) is the set of all possible outcomes of the experiment
(we always assume that it is not empty);
* F is an event space of subsets of Ω.
* P is a probability measure on F.
What are the 3 known properties of a probability space?
Let (Ω, F, P) be a probability space. Then, P has the following
properties
(a) If A, B ∈ F such that A ⊆ B, then
P(B − A) = P(B) − P(A).
Note that B − A = B ∩ A^c and is to be interpreted as ‘all elements of B that are not in A.
(b) For every A ∈ F,
P(A^c) = 1 − P(A).
(c) P(∅) = 0.
Proof.
(a) We write B as the union of the set with elements in B that are not in A and those that are, so B = (B − A) ∪ A. By finite additivity
P(B) = P((B − A) ∪ A) = P(B − A) + P(A), which proves the claim.
(b) Using the above property,
P(A^c) = P(Ω − A) = P(Ω) − P(A) = 1 − P(A).
(c) P(∅) = P(Ω^c) = 1 − P(Ω) = 1 − 1 = 0.
What is the Inclusion - Exclusion Formula?
Let (Ω, F, P) be a probability space. Then, for any finite collection A1, . . . , An of events in F, we have
P(Uk=1,n Ak) = Σk=1,n (−1)k−1Σ(1<=i1<···<ik<=n) P(Ai1 ∩ · · · ∩ Aik).
The second summation is embedded
Proof. We will prove the result only for n = 2 (the proof of the induction step in the general case is similar, but messier!). We write
A1 ∪ A2 = (A1 − B) ∪ B ∪ (A2 − B),
where B = A1 ∩ A2. The sets A1 − B, B, A2 − B are all disjoint, so we can write
P(A1 ∪ A2) = P ((A1 − B) ∪ B ∪ (A2 − B)) = P(A1 − B) + P(B) + P(A2 − B)
using finite additivity. We know by proposition 3.2 that P(A1 − B) = P(A1) − P(B) and similarly P(A2 − B) = P(A2) − P(B). Replacing these to the formula above, we get
P(A1 ∪ A2) = (P(A1)−P(B))+P(B)+ (P(A2)−P(B)) = P(A1)+P(A2)−P(B), which proves the claim.
What is Boole’s Inequality?
Let (Ω, F, P) be a probability space. If A1, . . . , An ∈ F, then
P(Ui=1,n (Ai)) <= Σi=1,n P(Ai) (*)
Proof. We proceed by induction. For n = 2, notice that
P(A1 ∪ A2) = P(A1) + P(A2) − P(A1 ∩ A2) <= P(A1) + P(A2).
so that () holds for n = 2. Assume now () holds ∀ j <= n, then we need to prove it holds for n + 1 events. Let A1, . . . , An+1 ∈ F, then arguing as above we have
P(Uj=1,n+1 Aj) = P((Uj=1,n Aj) U An+1)
= P((Uj=1,n Aj) + P(An+1) - P((Uj=1,n Aj) ∩ An+1)
<= P((Uj=1,n Aj) + P(An+1)
<= Σj=1,n P(Aj) + P(An+1) = Σj=1,n+1 P(Aj)
What is Conditional Probability?
Let (Ω, F, P) be a probability space and B ∈ F such that
P(B) > 0. For A ∈ F, the conditional probability of A given B is denoted by P(A|B) and is defined as
P(A|B) = P(A ∩ B)/P(B)
What is the Multiplication Rule of conditional probability?
(Multiplication Rule). Let (Ω, F, P) be a probability space and
A1, . . . , An ∈ F s.t. P(A1 ∩ · · · ∩ An−1) > 0. Then,
P(A1 ∩ · · · ∩ An) = P(A1)P(A2|A1)P(A3|A1 ∩ A2). . . P(An|A1 ∩ · · · ∩ An−1).
Proof.
Notice that for k = 1, . . . , n − 1, A1 ∩ · · · ∩ Ak ⊃ A1 ∩ · · · ∩ An−1. Hence, by Proposition 3.4 and by assumption
P(A1 ∩ A2 ∩ · · · ∩ Ak) > P(A1 ∩ · · · ∩ An−1) > 0 .
This ensures that all the conditional probabilities at the right hand side are welldefined. The result follows by a direct application of the definition of conditional probability on the right-hand-side:
P(A1)P(A2|A1)P(A3|A1 ∩ A2). . . P(An|A1 ∩ · · · ∩ An−1)
= P(A1) * P(A1 ∩ A2)/P(A1) * P(A3 ∩ A1 ∩ A2)/P(A1 ∩ A2) * . . .
* P(A1 ∩ · · · ∩ An)/P(A1 ∩ · · · ∩ An−1)
= P(A1 ∩ · · · ∩ An).
What is the Law of Total Probability?
Let (Ω, F, P) be a probability space and {Bn : n = 1, . . . , N} with N is either finite or infinite, be a partition of Ω
such that P(Bn) > 0, ∀ n = 1, . . . , N. Then, for all A ∈ F
P(A) = Σn=1,N [P(A|Bn)P(Bn)].
Proof. Notice that since {Bn : n = 1, . . . , N} forms a partition of Ω, we have A = A ∩ Ω = A ∩ Un=1,N Bn
= Un=1,N (A ∩ Bn).
Further, since Bn’s are disjoint so are {A ∩ Bn : n = 1 . . . N}, therefore by finite/countable additivity, we have
P(A) = P(Un=1,N (A ∩ Bn))
= Σn=1,N P(A ∩ Bn)
= Σn=1,N P(A|Bn)P(Bn).
In the last equality we use the definition of conditional probability with the assumption that P(Bn) > 0 ∀ n = 1, . . . , N.
What is Bayes’ Theorem?
Let (Ω, F, P) be a probability space and {Bn :
n = 1, . . . , N} with N is either finite or infinite, be a partition of Ω such that P(Bn) > 0 ∀ n = 1, . . . , N. Then for A ∈ F such that P(A) > 0
P(Bn|A) = P(A|Bn)P(Bn)/Σj=1,N P(A|Bj )P(Bj ) ∀ n = 1, . . . , N .
Proof. By definition of conditional probability and since A is such that P(A) > 0 then by the definition of conditional probability and the law of total probability:
P(Bn|A) = P(Bn ∩ A)/P(A)
= P(A|Bn)P(Bn)/P(A)
= P(A|Bn)P(Bn)/Σj=1,N P(A|Bj )P(Bj )
When are two events Independent?
Let (Ω, F, P) be a probability space. We say that events A and
B are independent if P(A ∩ B) = P(A) · P(B).
What does it mean for events to be pairwise and mutually independent?
Let (Ω, F, P) be a probability space and A1, A2, . . . , An be
events. We say that the events A1, . . . , An are pairwise independent if Aj and Ak are independent for every choice of j and k distinct. We say that the events A1, . . . , An are mutually independent, if
P(Aj1 ∩ Aj2 ∩ · · · ∩ Ajk) = P(Aj1)P(Aj2)· · · P(Ajk)
What is a Random Variable?
Let (Ω, F, P) be a probability space. A random variable is a function X : Ω → R such that {ω ∈ Ω : X(ω) <= a} ∈ F for every a ∈ R.
What is the Distribution of a random variable?
The distribution of a random variable X is the probability measure on R denoted PX and given by
PX(B) = P({ω ∈ Ω : X(ω) ∈ B})
for subsets B in some event space on the set of real numbers.
We also have that PX is a probability measure on R.
What is a Probability Mass Function?
Let (Ω, F, P) be a probability space and X be a discrete random variable. We define the probability mass function of X as the function pX : R → [0, 1] given by
pX(x) = PX({x}).
What is the Discrete Support?
Let (Ω, F, P) be a probability space and X be a discrete random variable. We define the discrete support of X, or, more precisely, the discrete support of its distribution PX, as the set
{x ∈ R : pX(x) > 0}.