1 basics Flashcards

1
Q

Let f : R to R
x_0 ∈R
consider
x’(t)=f(x(t)) for t ∈ [0,T]
x(0)=x_0

Give a
A sol

A

A sol is a func x: [0,T] to R of class C^1 and satisfies equalities

e.g. let r>0 x’(t) = rx(t) t ∈ [0,T]
x(0)=1

unique sol
x(t) = exp(rt) t ∈ [0,T]

linear differential equation:
solve by multiplying both sides by exp(-rt) and then noting its integrating by part form to solve

exp(-rt) x(t) = 1
x(t) = exp(rt)

chain rule and ftoc

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

exp growth

is brownian motion differentiable?

A

links to avg value of stock, but the value of stock is more stochastic in nature

Brownian motion/wiener process is an example of a stochastic process

brownian motion is non differentiable at “corners” nowhere differentiable, everywhere continuous

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

2d brownian motion

A

modelled in 2 components
W_1 and W_2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

A union B

A

{x: x∈A or x∈B}

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

A intersection B

A

{x: x∈A and x∈B}

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

A complement

A

Aᶜ:= Ω \ A ∈ F

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

sample space Ω

A

Non empty set of outcomes

subsets of the sample space Ω are EVENTS collection of all EVENTS is a SIGMA ALGEBRA

Ω COUNTABLE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

X={1,2,3}

P(X)
how many elements

A

P(X)=
{∅,{1,2,3}, {1},{2},{3}, {1.2}, {1,3}, {2,3}}

2^3 = 8 elements incl empty set

the power set is an example of sigma algebra and is the largest example for a given X

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Definition 1.1.1. A σ-algebra

A

Definition 1.1.1. A σ-algebra F on Ω is a collection of subsets of Ω such that
(i) Ω ∈ F (WHOLE SPACE IS AN ELEMENT)
(ii) If A ∈ F then Aᶜ:= Ω \ A ∈ F COMPLEMENTS

(iii) If Aₙ ∈ F for n ∈ N, then ∪ₙ₌₁ ∞ Aₙ ∈ F. (COUNTABLE UNIONS)

(EMPTY SET WILL BE IN)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

∅ ∈ F

F σ-algebra

A

(EMPTY SET WILL BE IN)
Let (Ω, F) be a measurable space. Since F is a σ-algebra which is closed under complementation Ωᶜ∈ F ∅ ∈ F

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

σ-algebra
closed under intersections?

Show that if Aₙ ∈ F for n ∈ N, then ∩ₙ₌₁ ∞ Aₙ ∈ F

A

If Aₙ ∈ F for n ∈ N then Aₙᶜ ∈ F

By De-Morgan’s identity
∩ₙ Aₙ = ∩ₙ (Aₙᶜ)ᶜ = (∪ₙ Aₙᶜ )ᶜ
which is in F because σ-algebras are closed under complementation and countable union.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

The pair (Ω, F)

A

The pair (Ω, F) is called a measurable space

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

measurable sets

A

elements of the sigma algebra

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

summary sigma algebra

A

(If i have an event A in sigma algebra I want the complement of my event also in)
(for two events A and B i want unions and intersections to be in the sigma algebra

sequences of complements and unions also contained

any sigma algebra on A is either {complement of A,A} or bigger up to P(A)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Exercise 1.1.3. Consider the set of real numbers R.
(1) Find the smallest σ-algebra on R.

A

{∅,R}
∅, R ∈ F by the def. of σ-algebras

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Exercise 1.1.3. Consider the set of real numbers R.

(2) Find the smallest σ-algebra on R that contains the interval (0, 1).

A

{∅,R, (0,1), (−∞, 0] ∪ [1, ∞)}

∅, R ∈ F by the def. of σ-algebras
since σ-algebras are closed under complementation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Def σ(A)
1.1.4

A

Let Ω be a non empty set and let A be a collection of subsets of Ω. We denote
by σ(A) the intersection of all σ-algebras on Ω that contain A, that is
σ(A) := ∩_{B∈Sₐ}B,
where Sₐ = {B : B is a σ-algebra on Ω and
A ⊂ B}.

This is a sigma algebra

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

The σ-algebra generated by A

A

σ(A)
intersection of all s.t they contain A
smallest σ-algebra that contains A.

A is a subset of the power set on Ω

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Exercise 1.1.5. Show that S_A is non empty.

S_A:= {B : B is a σ-algebra and A ⊂ B}

A

Let Ω be nonempty set and let A be a collection of subsets of Ω.

Let S_A:= {B : B is a σ-algebra and A ⊂ B}
Then P(Ω) ∈ S_A, because P(Ω) is a σ-algebra, and A ⊂ P(Ω). Hence S_A is non empty.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

ex 1.1.5 Show that intersection of σ-algebras is a σ-algebra. Consequently, σ(A) is indeed a σ-algebra.

A

Let H be a family of σ-algebras.
Then
1) by first axiom Ω ∈ F for all F ∈ H. Therefore
Ω ∈ ∩_{F∈H} F

2) Let A ∈ ∩_{F∈H} F

Then A ∈ F ∀F ∈ H.
So since σ-algebras are closed under complementation:
A^c ∈ F ∀F ∈ H.

Hence
A^c ∈ ∩_{F∈H}F.

So ∩_{F∈H} F is closed under complementation.

3) Let A_1,A_2,… ∈ ∩_{F∈H} F
Then
A_1,A_2,… ∈ F ∀F ∈ H.

So since σ-algebras are closed under countable union:
∪ₙ Aₙ ∈ F ∀F ∈ H.
Hence
∪ₙ Aₙ ∈ ∩{F∈H} F.
So ∩
{F∈H} F is closed under countable union.

By 1,2,3: ∩_{F∈H} F is a σ-algebra.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

e.g Ω ={1,2,3,4,5}

A={{1} {1,2}, {4,5}}
σ(A)

is there a sigma algebra F s.t A contained in F and if G is another sigma algebra s.t contains A then F contained in G

σ-algebra generated by A

A

{∅,{1,2,3,4,5},
{1} {1,2}, {4,5},
{2,3,4,5}, {3,4,5}, {1,2,3},
{1,2,4,5}}

{3} {1,4,5}

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

e.g Ω =R

A={[0,1]}
is there a sigma algebra F s.t A contained in F and if G is another sigma algebra s.t contains A then F contained in G

σ-algebra generated by A

A

{∅,R, (0,1), (−∞, 0] ∪ [1, ∞)}

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Definition 1.1.6. The Borel σ-algebra

A

The Borel σ-algebra on Rᵈ, denoted by B(Rᵈ) is the σ-algebra generated by the
collection of open sets of Rᵈ

“Consider a sigma algebra generated by A”

A ={u in R s.t U is open}

countable unions of open intervals

The borel sigma algebra is the smallest sigma algebra containing open sets

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Exercise 1.1.7. Show that B(R) contains all sets of the form (a, b), (a, b],[a, b), [a, b], (−∞, a],
(−∞, a), (a, +∞), [a, +∞), for a, b ∈ R. Moreover, show that the σ-algebra B(R) is generated
by any of these classes

A

A_1: collection of all open intervals on R

A_2 collection of closed intervals
A_3: collection of (−∞, a]
A_4: collection of (−∞, a)
A_5: collection of (a,+∞)
A_6: collection of [a,+∞)
A_7: collection of closed subsets of R

Then we can show the sigma algebra generated by open sets is the same as the one generated by all of these.

We show σ(A_1)= σ(A_2)……

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Method from lectures Exercise 1.1.7. Show that B(R) contains all sets of the form (a, b), (a, b],[a, b), [a, b], (−∞, a], (−∞, a), (a, +∞), [a, +∞), for a, b ∈ R. Moreover, show that the σ-algebra B(R) is generated by any of these classes
To show ** σ(A_1) = σ(A)** ** σ(A_1) ⊆ σ(A)** Take a ∈ A_1 (a will be an open interval) and is therefore an open set) thus a ∈ A. Thus A_1⊆A. By defn, σ(A) is the smallest σ-algebra containing A. A ∈ σ(A). A_1 ⊆A ⊆σ(A) Claim σ(A_1)⊆ σ(A), by defn A_1 is contained in A thus any sigma algebra containing A_1 will have this σ(A_1) contained. As σ(A) contains A_1 we have this true. ** σ(A) ⊆ σ(A_1)** Take an open set A ∈ A_1 This means A is a countable union of open intervals A= ∪ₙ,∞ (aₙ,bₙ) (aₙ,bₙ) is an open interval, belongs to A_1 and thus σ(A_1). Thus we have a collection of elements that belong to a sigma algebra, since σ(A_1) is a sigma algebra the union is an element of a sigma algebra. ∪ₙ,∞ (aₙ,bₙ)∈ σ(A_1) . Thus A ⊆ σ(A_1). Now we have a class contained in the sigma algebra, by defn σ(A_1) is the smallest such sigma algebra, thus σ(A⊆ σ(A_1)
26
Method in notes Exercise 1.1.7. Show that B(R) contains all sets of the form (a, b), (a, b],[a, b), [a, b], (−∞, a], (−∞, a), (a, +∞), [a, +∞), for a, b ∈ R. Moreover, show that the σ-algebra B(R) is generated by any of these classes
(i) Recall that B(R) is defined to be the σ-algebra generated by the collection of all open sets of R. Hence B(R) contains all open sets of R. Thus (a, b) ∈ B(R). (ii) (a, b] = ∩_n=1,∞ (a, b +1/n) ∈ B(R), since open sets are Borel, and σ-algebras are closed under countable intersection. (iii) [a, b) =∩_n=1,∞ (a − (1/n), b) ∈ B(R) by the same reasoning. (iv) [a, b] = ∩_{n=1} ∞ (a −(1/n), b + (1/n) ) ∈ B(R) by the same reasoning. (v) (−∞, a] = ∪_{n=1}∞ [−n, a] ∈ B(R) since (by (iv)) closed intervals are Borel and σ-algebras are closed under countable union. (vi) (−∞, a) = ∪_{n=1} ∞ (−n, a) ∈ B(R) since open sets are Borel and σ-algebras are closed under countable union. (vii) (a, ∞) =∪_{n=1} ∞ (a, n) ∈ B(R) by the same reasoning as in (vi). (viii) [a, ∞) = ∪_{n=1} ∞ [a, n] ∈ B(R) by the same reasoning as in (v). Hence B(R) contains all sets of the form(a, b), (a, b], [a, b), [a, b], (−∞, a], (−∞, a), (a, ∞), [a, ∞). We will now show that any of these classes generates B(R)
27
Method in notes Exercise 1.1.7. Show that B(R) contains all sets of the form (a, b), (a, b],[a, b), [a, b], (−∞, a], (−∞, a), (a, +∞), [a, +∞), for a, b ∈ R. Moreover, show that the σ-algebra B(R) is generated by any of these classes continued
(I) Since B(R) contains all sets of the form (a, b), and σ({(a, b) : a < b}) is the smallest σ-algebra containing all sets of the form (a, b), we have σ({(a, b) : a < b}) ⊂ B(R). Since every open set in R is a countable union of (disjoint) open intervals, and σ-algebras are closed under countable union, it follows that σ({(a, b) : a < b}) must contain all open sets in R. Since B(R) is the smallest σ-algebra containing all open sets of R, we have B(R) ⊂ σ({(a, b) : a < b}) Therefore we conclude that σ({(a, b) : a < b}) = B(R) (II) Similarly, as B(R) contains all sets of the form (a, b], and σ({(a, b] : a < b}) is the smallest σ-algebra containing all sets of the form (a, b], we have σ({(a, b] : a < b}) ⊂ B(R) Since (a, b) = ∪_n=1 ∞ (a, b − {b − a}/{2n} ] , every open interval is a countable union of intervals of the form (a, b]. So since σ-algebras are closed under countable union, all open intervals must be contained in σ({(a, b] : a < b}). So since σ({(a, b) : a < b}) = B(R) is the smallest σ-algebra containing all open intervals, we have B(R) ⊂ σ({(a, b] : a < b}). Hence we conclude that B(R) = σ({(a, b] : a < b}) (III) Similarly: σ({[a, b) : a < b}) ⊂ B(R) and since (a, b) = ∪_{n=1} ∞ [a + {b − a}/{2n} , b) we also have that B(R) ⊂ σ({[a, b) : a < b}). Therefore σ({[a, b) : a < b}) = B(R). (IV) Similarly: σ({[a, b] : a < b}) ⊂ B(R) and since (a, b) = ∪_{n=1} ∞ [a + (b − a)/(3n), b − {b − a}/3n] we also have B(R) ⊂ σ({[a, b] : a < b}) Therefore σ({[a, b] : a < b}) = B(R) (V) We have σ({(−∞, a] : a ∈ R}) ⊂ B(R). Furthermore (a, b] = (−∞, b] \ (−∞, a], so since σ-algebras are closed under taking differences of sets, σ({(−∞, a] : a ∈ R}) must contain all sets of the form (a, b]. So since σ({(a, b] : a < b}) = B(R) is the smallest σ-algebra containing all sets of the form (a, b], it follows that B(R) ⊂ σ({(−∞, a] : a ∈ R}). Hence we conclude that σ({(−∞, a] : a ∈ R}) = B(R) (VI) Similarly, σ({(−∞, a) : a ∈ R}) ⊂ B(R), furthermore [a, b) = (−∞, b) \ (∞, b) and thus σ({[a, b) : a < b}) ⊂ σ({(−∞, a) : a ∈ R}), i.e. B(R) ⊂ σ({(−∞, a) : a ∈ R}). So we conclude that B(R) = σ({(−∞, a) : a ∈ R}) (VII) Similarly, σ({(a, ∞) : a ∈ R}) ⊂ B(R), furthermore (a, b] = (a, ∞) \ (b, ∞) and thus σ({(a, b] : a < b}) ⊂ σ({(a, ∞) : a ∈ R}) i.e. B(R) ⊂ σ({(a, ∞) : a ∈ R}). So we conclude that σ({(a, ∞) : a ∈ R}) = B(R). (VIII) Similarly, σ({[a, ∞) : a ∈ R}) ⊂ B(R), furthermore [a, b) = [a, ∞) \ [b, ∞) and thus σ({[a, b) : a < b}) ⊂ σ({[a, ∞) : a ∈ R}) i.e. B(R) ⊂ σ({[a, ∞) : a ∈ R}). So we conclude that σ({[a, ∞) : a ∈ R}) = B(R). This finishes the proof. Note: In our proof we used the fact that σ-algebras are closed under taking differences of sets. This can be seen as follows: Let A, B be measurable sets. Then A \ B = A ∩ B^c is measurable, since σ-algebras are closed under complementation and (countable and thus also finite) intersections
28
σ-algebras are closed under taking differences of sets?
Note: In our proof we used the fact that σ-algebras are closed under taking differences of sets. This can be seen as follows: Let A, B be measurable sets. Then A \ B = A ∩ B^c is measurable, since σ-algebras are closed under complementation and (countable and thus also finite) intersections
29
Borel sigma algebra considered
On R^d we will always consider the Borel σ-algebra B(R^d). we choose whichever generates the set most convenient for us
30
Def 1.1.8 F/F' measurable
Let (Ω, F) and (Ω', F') be two measurable spaces. A function f : Ω → Ω' is called **F/F'-measurable** if f⁻¹(B) ∈ F for all B ∈ F' (If I take any set in F',B, the inverse image of this set under f is the set) f⁻¹(B) ={w∈Ω| f(w) ∈B}
31
e.g. suppose we have function f mapping: from c to Ω' ω₁→1 ω₂→3 ω₃→2 ω₄→1 f⁻¹({1,2}) = when is it measurable F/F' measurable
f⁻¹({1,2}) ={ω₁,ω₃,ω₄} For the function to be measurable we want this preimage to be a subset of the original Ω for every subset of Ω'
32
F-measurable. Ω' = R^d,
If Ω' = R^d, then as already mentioned we will always consider F' = B(R^d). Moreover, in this case, for a F/B(R^d)-measurable function we will simply say it is F-measurable.
33
e.g Let Ω={ω₁,ω₂,ω₃,ω₄} σ-algebra F ={∅, Ω, {ω₁},{ω₂,ω₃,ω₄}} what is this?
σ({ω₁}) ie the smallest sigma algebra generated by {ω₁}
34
Let Ω={ω₁,ω₂,ω₃,ω₄} σ-algebra F ={∅, Ω, {ω₁},{ω₂,ω₃,ω₄}} Let Ω'={1,2} σ-algebra F' ={∅, Ω, {1},{2}} define function f mapping Ω to Ω': ω₁→1 ω₂→2 ω₃→2 ω₄→2 check if F/F' measurable
Checking every set: f⁻¹(∅)= ∅ f⁻¹(Ω)= f⁻¹({1,2})= {w∈Ω| f(w) ∈Ω') } ={ω₁,ω₂,ω₃,ω₄}= Ω f⁻¹({1})= {ω₁} f⁻¹({2})= {ω₂,ω₃,ω₄} These we are elements in F thus the function f is measurable
35
Let Ω={ω₁,ω₂,ω₃,ω₄} σ-algebra F ={∅, Ω, {ω₁},{ω₂,ω₃,ω₄}} Let Ω'={1,2} σ-algebra F' ={∅, Ω, {1},{2}} Define function g mapping Ω to Ω': ω₁→1 ω₂→1 ω₃→2 ω₄→2 check if g is F/F' measurable
No, clearly always g⁻¹(∅)= ∅ g⁻¹(Ω)= f⁻¹({1,2})= {w∈Ω| f(w) ∈Ω') } ={ω₁,ω₂,ω₃,ω₄}= Ω g⁻¹({1})= {ω₁, ω₂} but doesn't belong to F not an element of F
36
Lemma 1.1.9 for two measurable spaces
Let (Ω, F) and (Ω', F') be two measurable spaces. Let A' be a collection of subsets of Ω' and assume that F' = σ(A'). (the sigma algebra is generated by a collection of subsets) Then, a function f : Ω → Ω' is F/F' -measurable **if and only if** f⁻¹(B) ∈ F for all B ∈ A' (A' the generating class) ------------------------------------------ if inverse image is a subset then is measurable but in this case F' is generated by class A so we don't need to check the inverse for all elements of F but only of B ∈ A', the generating class A' In particular If target space is R and borel sigma algebra, suffices to check inverse image of one of our prev A_i is a measurable set .
37
Exercise 1.1.10. Let (Ω₁, F₁), (Ω₂, F₂), and (Ω₃, F₃) be measurable spaces. Let f : Ω₁ → Ω₂ be F₁ /F₂-measurable and let g : Ω₂ → Ω₃ be F₂/F₃-measurable. Show that g ◦ f : Ω₁ → Ω₃ is F₁/F₃-measurable COMPOSITION
We know g ◦ f : Ω₁ → Ω₃ is defined on Ω₁.To show that g ◦ f : Ω₁ → Ω₃ is F₁/F₃-measurable: Let A∈F₃. We want to show that inverse image of A belongs in F₁. (g ◦ f)⁻¹(A) = f⁻¹(g⁻¹(A)). We know g is F₂/F₃-measurable. Thus g⁻¹(A)∈F₂. Since f is F₁ /F₂-measurable f⁻¹(g⁻¹(A))∈F₁. Thus the composition is F₁/F₃-measurable:
38
Exercise 1.1.11. Let f : R → R be a continuous function. Show that it is B(R)-measurable (Borel measurable)
Went through (for students with some background with analysis, "don't get too frustrated" We need to show that f is B(R)/B(R)-measurable. Since B(R) is generated by the collection of open sets in R, recalling Lemma 1.1.9 it suffices to show that f⁻¹(A) ∈ B(R) ∀A ⊂ R such that A is open. To this end, let A ⊂ R be open. Since f is continuous, the preimage f⁻¹(A) is also open. So since open sets are Borel, f⁻¹ (A) ∈ B(R), as required.
39
Consider f:R to R f(x)= 1 is this measurable?
(Ω, F) to (R, B(R)) the Borel sigma algebra is considered here on R To check measurable look at inverse image f⁻¹(A) for A ∈B(R) and check it is in F. Or from the lemma check if A in the form A=(a,b). f⁻¹(A) = {w∈Ω| f(w) ∈A } = {w∈Ω| 1 ∈A } = {Ω if 1 ∈A {∅ if 1 not in A constant functions are measurable (swap 1 to constant c)
40
Proporition 1.1.12 f+g
Let (Ω, F) be a measurable space and let f, g : Ω → R be F-measurable functions. Then f + g and f · g are F-measurable. (Here we are using borel as on R)
41
Exercise 1.1.13. CONSIDERING SEQUENCES OF FUNCTIONS Let (Ω, F) be a measurable space and let fₙ : Ω → R be F-measurable functions (borel measurable sequence of functs) for n ∈ N. inf fₙ are measurable.
Similarly, since the collection of sets of the form [a, ∞) generates B(R), it suffices to show that (infₙ fₙ)⁻¹ ([a, ∞)) ∈ F ∀a ∈ R. Now, (infₙ fₙ)⁻¹ ([a, ∞)) = {ω ∈ Ω : infₙ fₙ ∈ [a, ∞)} = {ω ∈ Ω : ∀n ∈ N, fₙ(ω) ≥ a} =∩_{n∈N}{ω ∈ Ω : fn(ω) ≥ a} =∩_{n∈N} fₙ⁻¹ ([a, ∞)) which is in F since [a, ∞) is Borel, fₙ is F/B(R)-measurable, and σ-algebras are closed under countable intersection. Hence infₙ fₙ is also measurable
42
Exercise 1.1.13. CONSIDERING SEQUENCES OF FUNCTIONS Let (Ω, F) be a measurable space and let fₙ : Ω → R be F-measurable functions (borel measurable sequence of functs) for n ∈ N. (1) Show that supₙ fₙ, Will be F measurable
𝒻~ caligraphic A I want to check the inverse image of a borel set is a borel set Checking : supₙ (fₙ⁻¹ (A)) From the lemma it suffices to check smaller class A∈B(R)= σ(𝒻) We choose A ⊂𝒻 most convenient 𝒻 = {(-∞,α]| α∈R} (supₙ fₙ)⁻¹ ((-∞,α]) = {ω ∈ Ω : supₙ fₙ(ω) ∈(-∞,α] } = {ω ∈ Ω : supₙ fₙ(ω) ≤ α } (sup< then all are less than) ={ω ∈ Ω : supₙ fₙ(ω) ≤ α for all n∈N } =∩ₙ {ω ∈ Ω : supₙ fₙ(ω) ≤ α } =∩ₙ fₙ⁻¹ ((-∞,α]) ((-∞,α]⊂B(R) is a borel set; the inverse image of a borel set via a measurable function is a measurable set so fₙ⁻¹ ((-∞,α])∈F for all n in N) Thus for a sequence of elements in the sigma algebra, the intersection will also be in ∩ₙ fₙ⁻¹ ((-∞,α]) ∈F thus measureable
43
Summary from exercises for sequences of measurable functions fₙ sup and inf
provided each fₙ is measurable supₙ fₙ is measurable and infₙ fₙ is measurable Thus the limit lim inf fₙ is measurable (limₙ inf fₙ = supₖ(inf _ₙ≥ₖ fₙ ) take the infimum of fₙ for n≥k looking at infimum of less and less things means its a increasing sequence in k for limits of sequences of fₙ is measurable functions and the limits exists then the limit is measurable function too limits, compisitions of measurable functions are measurable!
44
Exercise 1.1.13. CONSIDERING SEQUENCES OF FUNCTIONS Let (Ω, F) be a measurable space and let fₙ : Ω → R be F-measurable functions (borel measurable sequence of functs) for n ∈ N. (2) Show that lim supv→∞ fₙ is a measurable functionf
didnt go through in particular but mentioned it If for each ω ∈ Ω the limit limn→∞ fₙ(ω) exists, then f(ω) = limₙ→∞ fₙ(ω) = lim inf ₙ→∞ fₙ(ω) = sup_{n≥0} inf_{m≥n} fₘ(ω) which is measurable by part 1. of the exercise.
45
Def 1.1.1.4 PROBABILITY MEASURE
Let (Ω, F) be a measurable space. A non-negative function P : **F** → **[0, 1]** is called a **probability measure** if 1. P(∅) = 0 2. If Aₙ ∈ F, for n ∈ N, and Aₙ ∩ Aₘ = ∅ for n ≠ m, then P(∪ₙ₌₁∞ Aₙ) = Σₙ₌₁∞ P(Aₙ) (disjoint union = sums of individual probabilities) 3. P(Ω) = 1 (Funct P is defined for arguments which are sets)
46
(Ω, F, P)
PROBABILITY SPACE
47
A measure
A non-negative function µ : F → [0, +∞] satisfying 1 and 2 of the above definition will be called a measure A probability space is a particular case of a measure space
48
Let (Ω, F, P) be a probability space. A set A ⊂ Ω is called null, if
if there exists B ∈ F such that P(B) = 0 and A ⊂ B
49
The probability space (Ω, F, P) is called **complete** if
The probability space (Ω, F, P) is called complete if F contains all the null sets.
50
Exercise 1.1.15. Show that if A, B ∈ F and A ⊂ B, then P(A) ≤ P(B).
The probability measure gives properties for disjoint unions so rewrite as a disjoint set union We have that B= A∪ (B\A) Since A and (B \ A) are disjoint, we get P(B) = P(A) + P(B \ A). Since P(B \ A) ≥ 0, we get P(B) ≥ P(A).
51
Consider a probability space modelling rolling a die (Ω, F, P)
Ω= {1,2,3,4,5,6} F= P(Ω)={A| A⊆Ω} power set = {{1,2,3,4,5,6}, ∅, {1}, {2,3,4,5,6},....} 2^6 elements P(A)= #A/6 P({2,4,6}) = 3/6=1/2 take disjoint sets union prob= sum P( {2,3} U {5,6})= P ({2,3,5,6})= 4/6 or sum of them
52
Proposition 1.1.16 (Continuity from above and below). Limits of intersections and unions for certain sequences
(Ω, F, P) Let Aₙ, Bₙ ∈ F, for n ∈ N such that Aₙ ⊂ Aₙ₊₁ and Bₙ₊₁ ⊂ Bₙ for all n ∈ N. Then, i) P(∪ₙ Aₙ) = lim_{n→∞} P(Aₙ) Sequence of events which is increasing (largerset cont original set) ii) P(∩ₙ Bₙ) = lim{n→∞} P(Bₙ) Sequence of events which is decreasing (smaller and smaller subsets of a set)
53
Proposition 1.1.16 (Continuity from above and below). WHEN DOESNT IT WORK When we don’t have decreasing increasing sequence
In general not true, consider the real line Ω,=[0,1] P((a,b)) = b-a B_1= [1/2,1] B_n = [0,1] for all n >=2 then intersection of all B_n's is [1/2,1] which has measure 1/2 but the limit of the final probability should be one (its one for all B_n where n =2,3,4,.... thus this doesnt work as don't decrease
54
Exercise 1.1.17. Consider (Ω, F, P) Let A₁, A₂ be events such that P(A₁) = P(A₂) = 1. Show that P(A₁ ∩ A₂) = 1.
remark: A₁ and A₂ are not Ω; might have something smaller but not equal to Ω; there exists which have measure 0 without being the empty set e.g signleton set A={1/2} in measure [1/2 -e, 1/2 + e] = 2e will have measure 0} (We know as A₁⊆ A₁∩A₂ andA₂ ⊆ A₁∩A₂ the probabilities will be ≤) proof will differ from notes Sol: Let us prove that **P (A₁∪A₂)= P(A₁)+P(A₂)- P(A₁∩A₂)** We don't know if they are disjoint so we express as disjoint union A₁ ∪ A₂ = (A₁∩A₂) ∪ (A₁\A₂) ∪(A₂\A₁) Thus by the properties of prob measures with this sum of disjoint sets P(A₁∪A₂) = P(A₁∩A₂) +P(A₁\A₂)+P(A₂\A₁) (1) Write A₁ = (A₁\A₂) ∪ (A₁∩A₂) disjoint P(A₁) = P(A₁\A₂) + P(A₁∩A₂) Rearranges to P(A₁\A₂) = P(A₁)- P(A₁∩A₂) Write A₂ = (A₂\A₁) ∪ (A₁∩A₂) similarly rearranges to P (A₂\A₁) = P(A₂ )- P(A₁∩A₂) sub into (1) P(A₁∪A₂) = P(A₁∩A₂) +[ P(A₁)- P(A₁∩A₂)]+[ P(A₂ )- P(A₁∩A₂)] =P(A₁)+P(A₂)- P(A₁∩A₂) Thus by using this formula if P(A₁) = P(A₂) = 1 We know 1 ≥P(A₁∪A₂) ≥ P(A₁) =1 thus P(A₁∪A₂)=1 P(A₁∪A₂) = 2- P(A₁∩A₂) (≤ 1 (its a probability he didnt use that fact) Thus 1≤ P(A₁∩A₂). 1 = 2- P(A₁∩A₂) Thus P(A₁∩A₂) = 1
55
Exercise 1.1.18. Let Aₙ be events such that P(Aₙ) = 1 for all n ∈ N. Show that P(∩ₙ₌₁∞A) = 1. HInt: Use Proposition 1.1.16
events such that each have probability 1 their infinite intersection will also: Claim that ∀k∈N P( ∩ₙ₌₁ᵏ Aₙ) = 1 Base case: k=1 P( A₁ ) = 1 Induction hypothesis: Assume P( ∩ₙ₌₁ᵏ Aₙ) = 1 Inductive step: Consider P( ∩ₙ₌₁ᵏ⁺¹ Aₙ) = P( (∩ₙ₌₁ᵏ Aₙ) ∩ Aₖ₊₁ ) We know that P(Aₖ₊₁ )=1 by given sequence and also by induction hypothesis we have that P( ∩ₙ₌₁ᵏ Aₙ) = 1 And the previous exercise shows P( ∩ₙ₌₁ᵏ⁺¹ Aₙ) = P( (∩ₙ₌₁ᵏ Aₙ) ∩ Aₖ₊₁ ) = 1 Therefore we have shown by induction that this is true for all k in N. Dealing with infinity: Define new set Bₖ = ∩ₙ₌₁ᵏ Aₙ B₁ =A₁ B₂ = A₁∩A₂ ⊆ A₁= B₁ B₃= A₁∩A₂∩A₃ ⊆ A₁∩A₂ = B₂ In general Bₖ₊₁ ⊆ Bₖ decreasing seq of sets we know that for a decreasing seq the probability of the intersection is the limit of the probabilities ∩ₙ₌₁∞ Bₖ = ∩ₙ₌₁∞ ∩ₙ₌₁ᵏ Aₙ = ∩ₙ₌₁∞ Aₙ an intersection of decreasing sets thus P(∩ₙ₌₁∞ Aₙ) = P(∩ₙ₌₁∞ Bₖ) = lim_{ₖ to ∞} P(Bₖ) = 1 we showed by induction so limit is 1
56
Explaining {Aₖ infinitely open} {Aₖ i.o.} = ∩ₙ₌₁ ∞ ∪ⱼ₌ₙ∞ Aⱼ
e.g Aₙ ∪ Aₙ₊₁ ∪ Aₙ₊₂= Bₙ then take ∩ₙ₌₁ ∞ Bₙ Take w∈{Aₖ i.o.} ⟺ w∈ ∩ₙ₌₁ ∞ ∪ⱼ₌ₙ∞ Aⱼ ⟺ w∈ ∪ⱼ₌ₙ∞ Aⱼ for all n in N so there exists m>n s.t w∈ A_m w belongs to infinitely many A_n consider A₁,A₂,A₃,A₄,A₅,A₆,A₇,... Give me and n =4 say, i can find m >n=4 s.t w ∈A_m but if this is true for every n then I can always find a later set s.t w belongs, true for any n meaning w belongs to infinitely many.
57
Consider a coin flipped infinitely many times (Ω, F, P) Ω
Ω={H,T} flip once sample space Flip twice then sample space Ω={(H,H), (H,T), (T,H) ,(T,T)} NOTATION differs to {H,H} thus Ω_2 = { (a₁,a₂) | a_i∈{H,T}} Ωₙ = { (aₙ)ₙ₌₁∞ | a_i∈{H,T}} Ω = { (aᵢ)ᵢ₌₁∞ | aᵢ∈{H,T}} countable as its the # of reals
58
Consider a coin flipped infinitely many times (Ω, F, P) Define event
Ω = { (aᵢ)ᵢ₌₁∞ | aᵢ∈{H,T}} Lets consider set/event Aₙ= I get H nth time Aₙ={(a₁,a₂,a₃,..,aₙ₋₁Haₙ₊₁,...l) |aᵢ∈{H,T} for i ≠n} e.g A₂= {(H,H), (T,H)} = {(a₁,a₂)| |a₁∈{H,T} a₂ =H} ??? e.g A₂= {(H,H), (T,H), (T,H,T), (H,H,H), (H,H,T), (T,H,H),....} If I choose w ∈ Aₙ I know w= (a₁,a₂,a₃,..,aₙ₋₁H........) If we take w ∈ {Aₖ i.o.} ⟺ w∈ ∩ₙ₌₁ ∞ ∪ⱼ₌ₙ∞ Aⱼ ⟺ w∈ ∪ⱼ₌ₙ∞ Aⱼ for all n in N so there exists m>n s.t w∈ A_m w belongs to infinitely many A_n ie for n=1 there is an m_1 s.t w∈ A_m_1 n=2 there is an m_2 > m_1+1 s.t w∈ A_m_2.... So our H will be at position m_1,m_2,... true for any m thus we have infinitely many heads?
59
SUMMARY OF PROPERTIES OF A PROBABILITY SPACE
If we have a subset A of B then P(A)≤ P(B) B increasing sequence then intersection: i) P(∪ₙ Aₙ) = lim_{n→∞} P(Aₙ) Sequence of events which is increasing (largerset cont original set) ii) P(∩ₙ Bₙ) = lim{n→∞} P(Bₙ) Sequence of events which is decreasing (smaller and smaller subsets of a set)
60
Consider {Aₖ i.o.}ᶜ
complement of having infinitely many heads we will have finitely many heads we have heads up to some point, but after that there are only tails {Aₖ i.o.}ᶜ = (∩ₙ₌₁ ∞ ∪ₙ₌ₘ∞ Aₙ)ᶜ de Morgans twice gives = ∪ₙ₌₁ ∞ (∩ₙ₌ₘ∞ (Aₘᶜ)) Taking w∈ ∪ₙ₌₁ ∞ (∩ₙ₌ₘ∞ (Aₘᶜ)) = {Aₘᶜ eventually} thus there exists an n s.t w∈ ∩ₙ₌ₘ∞ (Aₘᶜ)) this means there exists n in N s.t for all m≥n w∈ Aₘᶜ ie for all positions greater than n we have w not in A_m so the position coords are not heads ie they are all tails
61
summary consider seq A_n then {A_n i.o} {A_n eventually}
then {Aₖ i.o.} = ∩ₙ₌₁ ∞ ∪ⱼ₌ₙ∞ Aⱼ {Aₙ eventually}= ∪ₙ₌₁ ∞ (∩ₙ₌ₘ∞ (Aₘᶜ))
62
Lemma 1.1.19 (Borel-Cantelli)
Consider (Ω, F, P) Aₖ⊆ F k in N Let (Aₖ)ₖ₌₁∞ be a sequence of events. (so they lie in our sigma algebra) If sumₖ₌₁∞ P(Aₖ) < ∞, then P(Aₖ i.o.) = 0, where {Aₖ infinitely open} {Aₖ i.o.} = ∩ₙ₌₁ ∞ ∪ⱼ₌ₙ∞ Aⱼ -------------------------------- If this is 0 then P({Aₖ i.o.}ᶜ)= P({ Aₖ eventually})= 1 the sum of the probabilities of the events {En} is finite then the probability that infinitely many of them occur is 0, that is,
63
finish prev lecture
12 feb 1.2 expectations? missed some?
64
65
Definition 1.1.20. random variable
Let (Ω, F, P) probability space function X:Ω → R is a **random variable* if X is F -measurable (F/B(R)- measurable) (meaning every time we take an element in B(R), then the pre-image of X is in F)
66
Exercise 1.1.21. Let X be a random variable on (Ω, F, P). Show that the collections of sets {X−1(A) : A ∈ B(R)} is a σ-algebra. In particular, it then follows that {X−1(A) : A ∈ B(R)} = σ({X−1(A) : A ∈ B(R)}). The σ-algebra above is called the σ-algebra generated by X and is denoted by σ(X).
notes
67
Remark 1.1.22. If we have two (or more) random variables X and Y then the σ-algebra generated by X and Y is
σ(X, Y ) := σ({X−1(A) : A ∈ B(R)} ∪ {Y −1(B) : B ∈ B(R)}). Notice that the collection {X−1(A) : A ∈ B(R)} ∪ {Y−1(B) : B ∈ B(R)} is not a σ-algebra in general(!). In order to ease the notation, most of the times that we deal with a random variable X we will drop the argument ω. For example, the set {ω ∈ Ω : X(ω) ≤ 1} will simply be denote by {X ≤ 1} of the probability P({ω ∈ Ω : X(ω) ≤ 1}) will simply be denoted by P(X ≤ 1)
68
* X induces a probability measure Q_X on R s.t Q_X: B(R) to [0,1]
Q_X is called the distribution of X or the law of X
69
Definition 1.1.23. DISTRIBUTION OF X
Let (Ω, F, P) be a probability space and let X : Ω → R be a random variable. The distribution of X is a probability measure Q_X on the (R, B(R)), which is given by Q_X(A) = P(X ∈ A) for all A ∈ B(R) . (=P(w∈Ω, X(w)∈A)) (if X is not measurable not defined)
70
elena side note: Distribution function of a R.V X
F_X(x)= P( X≤ x) =P(X in (-infinity,x]) is a particular case of Q_X remember those sets generate the whole B(R)
71
Definition 1.2.1 random var is SIMPLE
A random variable X is called simple if there exist c_1, ..., c_n ∈ R and events A_1, ..., A_n ∈F (sigma algebra) such that X =Σᵢ₌₁ⁿ cᵢ1_A_ᵢ
72
Expectation of X
If X is a **simple** with representation defined prev, then we define the **expectation of X** by E[X] : =Σᵢ₌₁ⁿ cᵢP(Aᵢ) It turns out that the expectation does not depend on the representation of X
73
Expectation is a linear function
if X, Y are simple random variables and a, b ∈ R, then E[aX + bY ] = aE[X] + bE[Y ].
74
Monotonicity of expectation
It is also easy to see that if X, Y are simple random variables eith X ≤ Y , then E[X] ≤ E[Y ].
75
If X is a non-negative random variable, then there exists
(If using that X is a non-negative RV then there exists a sequence random variables X_n, n ∈ N simple RVs such that for all ω ∈ Ω, X_n(ω) ↑ X(ω), as n ↑ ∞. Indeed, we can set X_n = h_n(X), where h_n(x) = min{[x2^n]2^{−n}, 2^n} where [a] denotes the lower integer part of a ∈ R.
76
Using that if X is a non negative RV then we can find an increasing sequence X_n of simple RVs s,t for all w∈Ω X_n (w) converges to ↑ X(w) as n converges to ↑infinity
Take X_n= h_n(X) where H_n(X)= min {x2^n} 2^{-n}, 2^n} (lower integer part?) (idea in notes)
77
Def: Expectation of X for any non-neg Rv
E(X) =ᵈᵉᶠ = limₙ→ ∞ E(X_n) where X_n is a sequence converging to X as n tends to infinity X_n is an increasing sequence well defined as X_n is a non decreasing seq
78
Def: RV X:Ω to R is **integrable** if
E[X+] < ∞, E[X−] < ∞, where X+=max{X,0} X-= max{-X,0} equivalently X is integrable IFF if E{|X|}< ∞
79
REMARK EEXPECTATIONS
Expectations are well-defined bc X+ X- and |X| are non-negative
80
L₁(Ω),
denote the set of all integrable random variables by L₁(Ω). L₁(Ω)={all integrable r.vs} ={X:Ω→R r.vs s.t E(|X|)< ∞ } AND For X ∈ L₁(Ω), we set E[X] = E[X⁺] − E[X⁻]
81
Lemma 1.2.2. Let X, Y ∈L₁(Ω),and a, b ∈ R. Then the following hold: The linearity and the monotonicity of the expectation is preserved on L₁(Ω)
Let X, Y ∈L₁(Ω),and a, b ∈ R. Then the following hold: 1. E[aX + bY ] = aE[X] + bE[Y ] linearity 2. If X ≤ Y , then E[X] ≤ E[Y ]. monotonicity
82
E(X) can also be denoted as
E(X)= Let X be a random variable. Then, for any non-negative B(R)-measurable g : R → R we have E[X)] =∫_Ω X dP or ∫_Ω X(w) dP(w) y the integral of X with respect to the measure P (P is a measure, lebesgue measure wrt to P)
83
Example expectations If X is a continous RV with PDF fₓ(x) Then what is E(X) E(g(x)
E(X)= ∫_R xfₓ(x) .dx E(g(X))= ∫_R g(x)fₓ(x) .dx can you always find pdf? For these cases we work with the law of distribution of the RV (for discrete is sum and probability mass funct)
84
Theorem 1.2.3. integral expectation for g
Let X be a random variable. Then, for any non-negative B(R)-measurable g : R → R we have E[g(X)] =∫_R g(x) Qₓ dx. where Qₓ is the distribution function (probably measure) valid for functs g taking values with different signs as long as well defined g+ and g- using g=g+ -g-
85
if X has a density function f, then E(g(X))
E(g(X))= ∫_R g(x)fₓ(x) .dx Actually, this is not a definition but a consequence of Theorem 1.2.3, since, the expression "X has a density f" means exactly that the distribution of X satisfies Qₓ(dx) = fₓ(x) dx.
86
||X||ₗₚ
= (E[|X|ᵖ])¹/ᵖ For p ∈ [1, ∞) (when p=1 this norm is the expectation of value X; collection of all random values in L_p space forms a linear subspace, 0 is constant therefore measurable and this tells you how far from **0**) (links to integral expectation E.G p=2)
87
||X||_L∞
= inf{K ≥ 0 : P(|X| ≤ K) = 1}. For p ∈ [1, ∞) so taking the K for which our RV absolute Probability is within 1
88
For p ∈ [1, ∞) denote Lₚ(Ω) L_p spaces
For p ∈ [1, ∞) denote Lₚ(Ω) the linear space of all random variables X : Ω → R such that ||X||ₗₚ < ∞. s,t the L_p norm is finite
89
Lemma 1.2.4 (Markov’s inequality).
Let X be a random variable. Then, for any ε > 0 we have P(|X| ≥ ε) ≤ (1/ε)E[|X|] probability greater than epsilon is bounded by a bound which depends on expectation
90
Lemma 1.2.5 (Hölders inequality).
Let p, q ∈ [1, ∞] with p⁻¹ + q⁻¹ = 1. Then for any X,Y rvs ||XY ||_L₁ ≤ ||X||_Lₚ ||Y ||_L_q L₁ norm is bounded for the product (generalisation of cauchys inequality) (E[|Xy|]) ≤ (E[|X|ᵖ])¹/ᵖ , (E[|Y|ᵠ])¹/ᵠ where power of q)
91
Lemma 1.2.6 (Minkowski’s inequality or triangle inequality)
Let p ∈ [1, ∞]. For any random variables X and Y we have ||X + Y ||_Lp ≤ ||X||_Lp + ||Y ||_Lp (L_p spaces are normed vector spaces) .
92
Exercise 1.2.7. Show that if p ≥ q, then ||X||_Lq ≤ ||X||_Lp
Hint: using Holders inequality Let p, q ∈ [1, ∞] with p⁻¹ + q⁻¹ = 1. Then for any X,Y rvs ||XY ||_L₁ ≤ ||X||_Lₚ ||Y ||_L_q Then ||1.X ||_L₁ ≤ ||1||_Lₚ ||X ||_L_q = (E[|1.X|]) ≤ (E[|1|ᵖ])¹/ᵖ . (E[|X|ᵠ])¹/ᵠ Suppose p ≥ q (ᵠ is ^q here) ||X||_Lq ᵠ = E[|X|^q] by defn = E[1 |X|^q] by holder's ≤ E[1][E[(|X|ᵠ)ᵖ/ᵠ]) ᵠ/ᵖ = ||X||_Lp ^q Hence, raising both sides of the inequality to the 1/q finishes the proof
93
Theorem 1.2.8 (Completeness of Lₚ (Ω)).
Let p ∈ [1, ∞]. Let X_n ∈ LLₚ (Ω), for n ∈ N such that lim_{n,m→∞} ||X_n − X_m||Lₚ = 0. (Cauchy) Then there exists X ∈ L(Ω), such that lim_n→∞ ||Xn − X||Lₚ = 0. (Every cauchy seq in the space converges to some value in the space)
94
1.3 Convergence of random variables
As we have seen, random variables are functions defined on a set Ω. As such, we have different modes of convergence.
95
Definition 1.3.1 (convergence almost surely)
For a sequence of RVs Let (Xₙ)_n∈N, be a sequence of random variables. We say that Xₙ converges to a random variable X almost surely if P{ω ∈ Ω : lim_{n→∞} Xₙ(ω) = X(ω)}= 1. We usually write Xₙ → X almost surely (This is a measurable set, we apply the probability measure to the measurable set)
96
Definition 1.3.2 (convergence in probability)
Let (Xₙ)_n∈N, be a sequence of random variables. We say that Xₙ converges to a random variable X in probability if for any ε > 0 lim {ₙ→∞} {ω ∈ Ω : |Xₙ(ω) − X(ω)| ≥ ε}= 0. We usually write Xₙ → X in probability. (probability: no matter how arbitrary small epsilon the limit)
97
Definition 1.3.3 (convergence in L_p).
Let (Xₙ)_n∈N, be a sequence of random variables that belong to ₚ(Ω). We say that Xn converges to a random variable X ∈Lₚ(Ω) in Lₚ if lim_{n→∞} ||Xₙ − X||_Lₚ = 0 We usually write Xₙ → X in Lₚ
98
Theorem 1.3.4. Relationship between types of convergence
Let (Xₙ)_ₙ∈N, X be random variables. The following hold. (i) If Xₙ → X almost surely, then Xₙ → X in probability. (ii) If Xₙ → X in probability, then there exists a subsequence (X_nₖ)k∈N such that X_nₖ → X almost surely. (iii) If Xₙ → X in Lp, then Xₙ→ X in probability.
99
Exercise 1.3.5. Show (iii) from Theorem 1.3.4 (iii) If Xₙ → X in L_p, then Xₙ→ X in probability
Let ε > 0. By Markov’s inequality, we have P(|Xₙ − X| > ε) = P(|Xₙ − X|ᵖ> εᵖ) ≤ (1/εᵖ) E[|Xₙ − X|ᵖ] = (1/εᵖ) ||Xₙ − X||ₗₚᵖ Taking limits: The right hand side of the above inequality tends to zero as n → ∞, by assumption. Hence, so does the left hand side. This shows that X_n → X in probability ignore: Let (Xₙ)_n∈N, be a sequence of random variables that belong to ₚ(Ω). Suppose Xₙ → X in Lₚ for X ∈Lₚ(Ω) in Lₚ THEN lim_{n→∞} ||Xₙ − X||_Lₚ = 0 by defn of convergence for all ε > 0 there is an N s.t for all n>N ||Xₙ − X||_Lₚ < ε thus... we need to show that for any ε > 0 lim {ₙ→∞} {ω ∈ Ω : |Xₙ(ω) − X(ω)| ≥ ε}= 0.
100
Theorem 1.4.1 (Monotone Convergence).
Let Xₙ, X : Ω → R, for n ∈ N, be non-negative random variables such that **almost surely** Xₙ ↑ X as n ↑ ∞. Then E[X] = limn→∞ E[Xₙ]
101
Lemma 1.4.2 (Fatou’s Lemma)
Let Xₙ : Ω → R, for n ∈ N, be non-negative random variables. Then E[lim inf n→∞ Xₙ] ≤ lim_n→∞ inf E[Xₙ]
102
Theorem 1.4.3 (Lebesgue’s theorem on Dominated Convergence).
Suppose that Xₙ → X **almost surely** and that **there exists** Y ∈ L₁(Ω) such that |Xₙ| ≤ Y . (bounded) Then X ∈ L_1(Ω) and E[X] = lim_n→∞ E[Xₙ].
103
Definition 1.5.1. The events A1, ..., An are called independent if
consider a probability space (Ω, F, P) The events A₁, ..., An are called independent if P(∩ ᵢ₌₁ ⁿ Aᵢ₌₁) = ∏ᵢ₌₁ ⁿ P(A ᵢ). probability is multiplies for intersections of indep events
104
sigma algebras G_1, ..., G_n are called independent if
sigma algebras G_1, ..., G_n are called independent if for any A_1 ∈ G1, ..., A_n ∈ G_n, the events A_1, ..., A_n are independent.
105
Random vars X_1, ...X_n are independent if
The random variables X_1, ..., X_n are independent if the σ-algebras σ(X_1), ...σ(X_n) are independent
106
the elements of a family of events, σ-algebras, random variables are independent the element of any finite sub-family are independent.
the elements of a family of events, σ-algebras, random variables are independent if the element of any finite sub-family are independent.
107
Lemma 1.5.2. independence
Let G_1, G_2 ⊂ F be σ-algebras. Let curly_A_1 ⊂ G_1 and curly_A_2 ⊂ G_2 be π-systems (classes that are closed under finite intersection) such that σ(curly_A_1) = G_1 and σ(curly_A_2) = G_2. (sigma algebras gen by these events) Then G_1 and G_2 are independent if and only if P(A_1 ∩ A_2) = P(A_1)P(A_2) for all A_1 ∈ curly_A1, A2 ∈ curly_A2.
108
Theorem 1.5.3. If X, Y ∈ L_1(Ω) are **independent random variables** then E(XY)
XY ∈ L_1(Ω) and E[XY ] = E[X]E[Y ]
109
Lemma 1.5.4. independence and measurable functs
Let X₁, ..., Xₙ be **independent** random variables, m₁ + ... + mₖ = n and f₁, ..., fₖ be measurable functions of Rᵐ₁, ..., R ᵐₖ respectively. Then the random variables Y₁ = f₁(X₁, ..., Xₘ₁), Y₂ = f₂(Xₘ_₁ ₊₁ , ..., Xₘ_₁ ₊ₘ_₂ ), ..., Yk = f(Xₘ_₁+...+ₘ_{ₖ−₁}₊₁, ..., Xₙ) are independent.
110
Definition 1.5.5 sub algebra existence with conditional expectation
Let X ∈ L_1(Ω) and let G ⊂ F be a sub-σ-algebra of F. One can show that there exists a unique Y ∈ L1(Ω) with the following two properties: 1. Y is G-measurable 2. E[**1**_A X] = E[**1**_A Y ] for all A ∈ G (indicator funct for A) Y ~ conditional expectation of X given G Y = E[X|G].
111
conditional expectation recap
For (Ω, F, P) and RV X X:Ω to R Let X ∈ L_1(Ω) ( IFF E[|x|] < infinity) and let 𝒢 ⊂ F be a sub-σ-algebra of F. There exists a unique RV Z∈ L_1(Ω) (z dep on Z and 𝒢) s,t E[**1**_A X] = E**1**_A Z] for all A ∈𝒢 So **conditional expectations** given a sigma algebra are a random variable themselves Z= E[X|𝒢]
112
Exercise 1.5.6. Suppose that Ω = ∪ᵢ₌₁ ⁿ Aᵢ where Aᵢ are disjoint events with P(Aᵢ) > 0. (DISJOINT PARTITION of sample space) Let 𝒢 = σ({A₁, ...., Aₙ}) (sigma algebra generated by a collection of subsets of Ω) What do elements look like?
elements of sigma algebra 𝒢 Aᵢ∈𝒢 unions ∪_{i in I} Aᵢ ∈𝒢 complements e.g complement of A_1 will be a union of other partition elements intersections and complements of these turn out to be unions turns out all elments are unions of the A_is
113
Exercise 1.5.6. Suppose that Ω = ∪ᵢ₌₁ ⁿ Aᵢ where Aᵢ are disjoint events with P(Aᵢ) > 0. (DISJOINT PARTITION of sample space) Let 𝒢 = σ({A₁, ...., Aₙ}) (sigma algebra generated by a collection of subsets of Ω) show that E[X|𝒢] is a simple random variable given by E[X|𝒢] = Σᵢ₌₁ ⁿ cᵢ **1**_Aᵢ where the constants cᵢ are given by cᵢ = (1/P(Aᵢ)) E[**1**_Aᵢ X]. (It follows from the above that if 𝒢= {∅, Ω} is the trivial σ-algebra, then E[X|𝒢] = E[X]. We list below some properties of conditional expectations
We want to show that almost surely on Ω, we have E(X|𝒢) = Y , i.e. that E[**1**_G X] = E[**1**_G Y ] ∀G ∈ 𝒢 We want to show that for all A∈𝒢 E[ **1**_A x]= E[ **1**_A Σᵢ₌₁ ⁿ cᵢ 1_Aᵢ] Suppose A= Aⱼ j∈{1,...,n} Then E[ **1**_Aⱼ Σᵢ₌₁ ⁿ cᵢ **1**_Aᵢ] =E[ Σᵢ₌₁ ⁿ cᵢ **1**_Aᵢ **1**_Aⱼ ] has value 1 only if we are considering the intersection, both 1 = E[ Σᵢ₌₁ ⁿ cᵢ **1**_Aᵢ ∩_Aⱼ ] = E[cⱼ **1**_Aⱼ ] =(***) ------------------------------- since A₁, . . . , Aₙ are disjoint, so **1**_Aᵢ ∩_Aⱼ = {1_∅ if i ≠ j (empty set has no elements) {1Aᵢ if i = j = {0 if i ≠ j {1_Aⱼ if i = j. ------------------------------------ Replacing cⱼ (***)= = E[cⱼ **1**_Aⱼ ] = E[(1/P(Aⱼ)) E[**1**_Aⱼ X] **1**_Aⱼ ] = cⱼ E[**1**_Aⱼ ] by linearity of expectation, multiplying by RV =(1/P(Aⱼ))) E[**1**_Aⱼ X] E[**1**_Aⱼ] by expectations of simple RVs (E[**1**_Aⱼ] =P(Aⱼ) =E[**1**_Aⱼ X] ------------- Now checking the case for all A∈𝒢 : Suppose A= ∪_{j∈I} Aⱼ j∈{1,...,n} E[ **1**_∪_{j∈I}Aⱼ Σᵢ₌₁ ⁿ cᵢ **1**_Aᵢ] = E[ Σⱼ **1**_Aⱼ Σᵢ₌₁ ⁿ cᵢ **1**_Aᵢ] expectation of sum =Σⱼ∈I E[**1**_Aⱼ Σᵢ₌₁ ⁿ cᵢ **1**_Aᵢ] singular A_j =Σⱼ∈I E[**1**_Aⱼ Σᵢ₌₁ ⁿ cᵢ **1**_Aᵢ] =Σⱼ∈I E[**1**_Aⱼ X] =E[**1**_{∪_{j∈I}Aⱼ} C ] -------------- characteristic of a union **1**_∪_{j∈I} =Σⱼ **1**_Aⱼ -------
114
suppose that Ω = ∪ᵢ₌₁ ⁿ Aᵢ where Aᵢ are disjoint events with P(Aᵢ) > 0. (DISJOINT PARTITION of sample space) since A₁, . . . , Aₙ are disjoint, so **1**_Aᵢ ∩_Aⱼ =
since A₁, . . . , Aₙ are disjoint, so **1**_Aᵢ ∩_Aⱼ = {1_∅ if i ≠ j (empty set has no elements) {1Aᵢ if i = j = {0 if i ≠ j {1_Aⱼ if i = j.
115
E[ Σᵢ₌₁ ⁿ **1**_Aᵢ] for Aᵢ disjoint "this is primary school question" E[ Σᵢ₌₁ ⁿ cᵢ **1**_Aᵢ] for Aᵢ disjoint
E[ Σᵢ₌₁ ⁿ **1**_Aᵢ] =Σᵢ₌₁ ⁿ P(Aᵢ) E[ Σᵢ₌₁ ⁿ cᵢ **1**_Aᵢ] =Σᵢ₌₁ ⁿ cᵢ P(Aᵢ) takes finitely many values
116
characteristic of a union **1**_∪_{j∈I}Aⱼ disjoint sets suppose that Ω = ∪ᵢ₌₁ ⁿ Aᵢ where Aᵢ are disjoint events with P(Aᵢ) > 0. (DISJOINT PARTITION of sample space)
**1**_∪_{j∈I}Aⱼ =Σⱼ **1**_Aⱼ = beacuse w belongs to exactly one of them, disjoint thus others are 0
117
in general are conditional expectations RVS?
In general conditional expectations are random variables not # they can be numbers if sigma algebra is trivial
118
Proposition 1.5.7. Let X, Y ∈ L1(Ω) and G ⊂ F be a σ-algebra. The following hold properties of conditional exp
Let X, Y ∈ L₁(Ω) (by defn can take the cond exp) and 𝒢 ⊂ F be a σ-algebra. (by defn its a RV so is F-measurable) The following hold 1. **Linearity**: If a, b are constants, then E[aX + bY |𝒢] = aE[X|𝒢] + bE[Y |𝒢]. 2. Taking out what is known: If X is 𝒢-measurable and XY is integrable, then E[XY |𝒢] = XE[Y |𝒢]. **In particular, if X is G-measurable, then E[X|𝒢] = X.** 3. **Conditioning** on independent information: If W ⊂ F is a σ-algebra independent of σ(σ(X), 𝒢), then E[X|σ(G, W)] = E[X|𝒢]. In particular, if σ(X) is independent of W, then E[X|W] = E[X]. 4. **Tower property**: If W ⊂ 𝒢 is a σ-algebra, then we have E[X|W] = E[E[X|𝒢]|W]. 5. **Monotonicity**: If X ≥ 0 then E[X|𝒢] ≥ 0. 6. **Conditional Jensen’s inequality**: If φ : R → R is convex, such that φ(X) ∈ L₁(Ω), then φ(E[X|𝒢]) ≤ E[φ(X)|𝒢].
119
Exercise 1.5.8. Show that the conditional expectation is a continuous operator from L₁(Ω) to L₁(Ω). In other words, show that if Xₙ, X ∈ L₁(Ω) such that lim_n→∞ ||Xₙ − X||_L₁ = 0, then lim_{n→∞} ||E[Xₙ|𝒢] − E[X|𝒢]||_L₁ = 0.
sequence of RVS converge in L₁ then conditional exp converges in L₁ Xₙ converges to X E[Xₙ|G] converges to E[X|G]| ------------- Let (Ω, F, P) be a probability space, and let 𝒢 ⊂ F be a sub-σ-algebra. Suppose that Xₙ, X ∈ L_1(Ω, F, P) and that limn→∞ ||Xₙ − X||_L₁Ω = 0. Then lim_n→∞ ||E[Xₙ|𝒢] − E[X|𝒢]||L₁(Ω) = as E[·|𝒢] is a linear operator: =lim_n→∞ ||E[Xₙ-X|𝒢]||_L₁Ω By L_1-norm abs value =lim_n→∞ E[ | E[Xₙ-X|𝒢] | ] since | · | is convex, by Jensen’s inequality: (absolute value inside) ≤ lim_n→∞ E[ E[ | Xₙ-X | |𝒢] ] by the law of total expectation(or tower property) lim_n→∞ E[ | Xₙ-X | ] by the def. of the L1(Ω)-norm: lim_n→∞ || Xₙ-X ||L₁(Ω) by assumption =0
120
note that E[ X |𝒢]
E[ X |𝒢] =E[X] ------------ suppose we take a trivial sigma algebra W= {Ø,Ω} Then E[ X |W] =E[X] It suffices to check that E[**1**_A E[X]]= E[**1**_AX] for all A ∈W sigma algebra W if this is true then LHS = E[ X |W] is true as can either by empty set or Ω easy to check for 2 tower property (E[E[X |𝒢] = E[ E[X |𝒢] |W] = E[ X |W]= E[X] , as the trivial sigma algebra is always contained in 𝒢
121
Exercise 1.5.9. Let X ∈ L₂(Ω) (meaning ||X||_L₂ < infinity) and let 𝒢 ⊂ F be a sub-σ-algebra of F. Show that ||X − E[X|𝒢]||_L₂ = min_Z∈L₂ᴳ ||X − Z||_L₂ where we have denoted by L₂ᴳ the collections of all 𝒢-measurable random variables that belong in L₂(Ω). L₂ᴳ subspace of L₂(Ω).
By defn the conditional expectation E[X|G] is G-measurable E[X|G] ∈L₂ᴳ(Ω). (check: E[X|G] is G-measurable and finite uder L2 norm) Moreover, by the conditional Jensen’s inequality we see that E[|E[X|G]|²] ≤ E[E[|X|²|G]] = E[|X|²] < ∞. Consequently, E[X|G] ∈ L₂ᴳ and by definition of the minimum we get ||X − E[X|G]||_L₂ ≥ min_{Z∈L₂ᴳ} ||X − Z||_L₂ (by definition I am taking the smallest from the set, for each RV Z i have a real # each one is above min) tbc .
122
prove equality by proving inequality in both directions ||X − E[X|𝒢]||_L₂ = min_Z∈L₂ᴳ ||X − Z||_L₂ proving ||X − E[X|G]||_L₂ ≤ min_{Z∈L₂ᴳ} ||X − Z||_L₂ part
On the other hand, for any Z ∈ L₂ᴳ ||X - Z||ₗ₂ ² = ||X - Z − E[X|G] + E[X|G] ||ₗ₂ ² By expanding ² = ||X − E[X|G]||ₗ₂ ² + E[(X − E[X|G])(E[X|G] − Z)] +||X - Z||ₗ₂ ² ≥ ||X − E[X|G]||ₗ₂² + E[(X − E[X|G])(E[X|G] − Z)] We claim the last term on the RHS is zero. Indeed, since (E[X|G] − Z) and E[X|G] are G-measurable, we have E[(X − E[X|G])(E[X|G] − Z)]= we can take an extra conditional expectation inside the expectation bc the expectation of the conditional expectation is the expection itself = E[E[(X − E[X|G])**(E[X|G] − Z)**|G]] "(E[X|G] − Z)is G measurable something times something g-measurable given g we can use the property to pull it out" = **E[(E[X|G] − Z)**E[(X − E[X|G])|G]] "claim (E[X|G] − E[X|G])] is 0,as conditional exp are linear , by defn conditional exp is G measurable so cond exp wrt g again doesnt change" = E[(E[X|G] − Z)(E[X|G] − E[X|G])] = E[(E[X|G] − Z) · 0] = 0. showing for any such Z the inequality holds
123
L₂(Ω)= L₂ᴳ (Ω)= L₂ᴳ subspace of L₂(Ω). where we have denoted by L₂ᴳ the collections of all 𝒢-measurable random variables that belong in L₂(Ω).
L₂ᴳ (Ω)= {X: Ω→R| X is 𝒢-measurable , ||X||_L₂ < ∞.} L₂(Ω)= {X: Ω→R| X is F-measurable , ||X||_L₂ < ∞.} L₂ᴳ (Ω)⊆L₂(Ω) this is because if I choose an X in L₂ᴳ (Ω) it is 𝒢-measurable thus inverse images of borel sets belong to 𝒢: ∀A∈B(R), X⁻¹(A)∈𝒢⊆F if I take any 𝒢-measurable it is automatically F-measurable due to 𝒢 ⊂ F being a sub-σ-algebra of F.
124
geometric interpretation L₂(Ω)= {X: Ω→R| X is F-measurable , ||X||_L₂ < ∞.}
consider R^2 plane taking vector in R^2 **x,y** in R^2 also operation +:closed under + **x+y** also in R^2 closed under scalar multiplication: a**x** in R ^2 if X is F-measurable , ||X||_L₂ < ∞: im claiming that for two RVS X,Y X+Y is also a RV in L_2 F-measurable ||X+Y||_L₂ by triangle =< ||X||_L₂ +||Y||_L₂ < ∞ is also finite aX is also a RV ||aX||_L₂ < a||X||_L₂ < ∞ another interpretation is thatL₂ᴳ (Ω)⊆L₂(Ω) lies on the line, any Z in L₂ᴳ (Ω) lies along the line, we can scale etc and maintain closure under these operations
125
scaling mean and variance