Chapter 1: Geometry of the Euclidean space Flashcards
R ⁿ is
the set of all n-tuples of real numbers,
Rⁿ =
{(x_1,…, x_n) : x_1, . . , x_n∈R}.
components are coordinates
natural dot product/inner product/scalar
product
x · y = Σ_i=1 ,n x_iy_i,
where x = (x_1, … , x_n), y = (y_1,. .., y_n) ∈ Rⁿ
Properties:
(i) (α₁x₁ + α₂x₂) · y
= α₁(x₁ · y) + α₂(x₂ · y)
for any x₁, x₂, and y ∈ Rⁿ, and any real numbers α₁, α₂ ∈ R; distributive linear in the first var
(ii) x·y = y·x for any x and y ∈ Rⁿ symmetry
(iii) x·x > 0 for any x ∈ Rⁿ, and
x · x = 0 if and only if x = 0.
positively homogeneous
properties reflect the structure in Rⁿ
Proposition I.1
Cauchy-Schwarz inequality
For any vectors x, y ∈ Rⁿ the following inequality holds:
|x · y| ≤
|x| |y|,
(where
|x| =√{x · x}, and
|y| =√{y · y} are lengths of vectors in Rⁿ)
Equality occurs IFF x = 0 or y = λx for some λ ∈ R.
Cauchy-Schwarz inequality alt ways of writing
|x · y| ≤
|x| |y|,
from i=1 to n
Σxᵢ · yᵢ ≤ (Σxᵢ² Σyᵢ² )¹/²
when n=1
x₁· y₁ ≤(x₁² · y₁² )¹/² = x₁· y₁ equality
abs value= length here
|x + y|≤|x|+|y|.
Proposition I.1
Cauchy-Schwarz inequality
PROOF L5
x=0 TRIVIAL CASE holds
|x| =√{x · x}
|λx-y|=√{λx-y · λx-y}
|λx-y|²={λx-y · λx-y}
assume x≠0
Consider polynomial p(λ)
= |λx-y|²
now using the properties of dot prod:
linearity, symmetry
=|λ²x² +y² -2λxy|
= λ²|x|² +|y|² -2λxy≥0
parabola in λ with at most one root
thus DISCRIMINANT ≤0
iff
(-2xy)² -4|x|²|y|² ≤0
4(xy)²-|x|²|y|² ≤0
thus inequality proved
Proposition I.1
Cauchy-Schwarz inequality
PROOF L5 CASE OF EQUALITY
Looking at the case of equality
IF CASE x=0 or y =λx (colinear)
|0|≤0
|λx * λx|= λ²|x*x|= λ²|x| |x|
only if case
suppose that |xy|≤ |x||y| for some x and y
if x≠0 then we need to show that y=λx
Considering the poly: then discriminant =0
thus only one real root and hence
λ₀ that is
0=P(λ₀ ) = (λ₀x -y)² length of vector is 0 by non-neg property of dot prod
iff
λ₀x -y=0
iff
y=λ₀x
L5 remark cauchy schwarz
Note that the argument used in the proof above is rather general and works for any (not
necessarily finite-dimensional) vector space V equipped with a scalar product <.,.>
L5:
scalar product <·, ·>
bilinear map V × V → R that satisfies the following properties, which mimic the properties of the dot product:
(i) <α_1u_1 +α2_u_2, v> = α_1<u_1, v>+α_2<u_2, v_i> for all u_1, u_2, and v ∈ V , and all reals α_1, α_2 ∈ R; LINEAR IN FIRST ARGUMENT
(ii) <u, v> = <v, u> for all u and v ∈ V ; SYMMETRY
(iii) <u, u> > 0 for any u ∈ V , and <u, u> = 0 if and only if u = 0
NON NEGATIVITY
Example:
Let V=C[a,b] be the space formed by all continuous vector functions f:[a,b] to R^n
is this a vector space
clearly, closed under linear operations (and f(x)=0?)
Scalar product on C[a,b]
<f,g>
<f,g> =
∫ₐᵇ f(t)g(t) .dt
Here our vector space is infinite dimension we can show it satisfies properties of the scalar properties
Lemma I.2 (Integral Cauchy-Schwarz inequality)
For any continuous vector functions f,
g : [a, b] → R^n the following inequality holds:
|∫ₐᵇ (f.g)(t) .dt| ≤
(∫ₐᵇ |f(t)|² .dt) ¹/² (∫ₐᵇ |g(t)|² .dt) ¹/²
The equality occurs if and only if f ≡ 0 or g ≡ λf for some λ ∈ R. (identically proportional?)
|<f,g>| ≤ sqrt(<f,f> <g,g>) = |f||g|
Proof: Lemma I.2 (Integral Cauchy-Schwarz inequality)
Let V be a vector space of all continuous vector-functions f : [a, b] → R^n
On this vector space we consider the following scalar product
<f,g> =ᵈᵉᶠ= ∫ₐᵇ (f.g)(t) .dt where f, g ∈ V
and the function in the integral is obtained by taking the dot product of the values f(t) and g(t). It is straightforward to check that the formula above indeed defines the scalar
product on V , that is it satisfies the properties (i)-(iii) above. Now the proof of the lemma follows the same line argument as in the proof of Lemma I.1, by considering the polynomial
P(λ) = <λf − g, λf − g>.
Corollary I.2 (Triangle inequality). Of CS
For any vectors x, y ∈ Rⁿ the following inequality holds:
|x + y|≤|x|+|y|.
Equality occurs IFF x = 0 or y = λx for some λ> 0 in R
(proportional/colinear for equality)
Corollary I.2 (Triangle inequality consequence of C-S).
PROOF
For any vectors x, y ∈ Rⁿ the following inequality holds:
|x + y|≤|x|+|y|.
Equality occurs IFF x = 0 or y = λx for some λ> 0 in R
proof:
By linearity and symmetry defn using values and lengths
|x + y|²
= (x + y) · (x + y)
= |x|² + x · y + y · x + |y|²
= |x|² + 2x · y + |y|²
≤|x|² + 2 |x · y| + |y|²
≤|x|² + 2 |x| |y| + |y|²
= (|x| + |y|)²
(second inequality used the Cauchy-Schwarz inequality)
Equality in the triangle inequality implies equality in the Cauchy-Schwarz inequality, and hence, implies that x = 0 or y = λx for some λ > 0.
shown in lecture
(if x=0 |y|²=|y|², if y= λx LHS:
|x + λx|= |(1+λ)x|= sqrt((1+λ)x·(1+λ)x )
RHS: |x|+ |λx| = |1+λ| |x| shows equality)
Conversely, if x = 0 or y = λx, where λ > 0, then the triangle inequality becomes an equality.
Suppose |x+y|² =( |x|+|y|)²
Use modulus and dot product relationship
(x+y) · (x+y)
= ( |x|+|y|)²
by cauchy schwarz conclude x=0 or y=λx
x not eq 0 (1+λ)² |x|² = (1-λ)²) |x|² if λ<0
however plugging y= λx into inequality |x+y| <= |x| + |y| we rule out the case when λ<0
distances in R^n
dist(x,y)
=|x−y|= (x−y)·(x−y)
From MATH2051 we also know that the dot product allows us to compute lengths of curves in R^n
Distances in R^n
dist(x,y)=|x−y|= (x−y)·(x−y).
From MATH2051 we also know that the dot product allows us to compute lengths of curves in R^n
Triangle inequality in form for distances
In particular, the triangle inequality can be written in the following form
for any vectors x, y, and z ∈ Rn. (Make sure that you can explain why.)
dist(x, y) less than it equal to
dist(x, z) + dist(z, y)
|x-y| inequality for triangle
less than or equal to |x-z| + |z-y|
iff
|u+v| <= |u| +|v|
u=x-z and v= z-y
x·y./ (|x| |y|)
x·y./ (|x| |y|) in [-1,1]
hence choose cos theta
also choosing thete in [0, pi]
assuming x,y non zero vectors
to compute the cosine of angles
θ
between non-zero vectors x and y
by the formula
cosθ= x·y./ (|x| |y|)
RHS cos θ for some θ is a consequence of the Cauchy-Schwarz inequality: it guarantees that the quotient on the right hand-side above takes values in the interval [−1, 1].
The equality case in the Cauchy-Schwarz inequality says that the angle θ between non-zero vectors x and y equals πk, where k ∈ Z, if and only if the vectors are collinear.
Recall that vectors x and y are called orthogonal if x · y = 0, i.e. the angle between them equals π/2 + πk, k ∈ Z.
Pythagorean theorem:
vectors x and y are orthogonal if and only if |x + y|^2 = |x|^2 + |y|^2.
Particular case of
Distance using Cauchy Schwartz
. to compute volumes of parallelotopes (not covered in lecture but was in exercise !)
in more detail, for a system e1, . . . , ek of k linearly independent vectors the k-dimensional parallelotope Pk, spanned by them, is defined as
Pk ={t1e1 +…+tkek :ti ∈[0,1]}. Its k-dimensional volume is given by the formula
Vol_k (P_k) =
√[
Det(
[ e ₁. e₁ e ₁.e₂ …. e₁.eₖ]
…..
[ eₖ. e₁ eₖ. e₂ …. eₖ.eₖ])
In particular, if the ei’s are pair-wise orthogonal (that is Pk is an orthotope ), then we obtain
Volk(P_k)= SQRT((e1 ·e1)···(ek ·ek) )=|e_1|…..|e_k|
def 1.2 vector subspace’
linear vector space
A subset X ⊂ R^n st for any x, y ∈ X and any a, b ∈ R we have ax + by ∈ X.
closed under linear combos,
subspace commonly solutions to linear equations
0 is always an element of a vector subspace
we use properties of dot prod and vector space properties
we define a subspace as a span of a collection of vectors min # give basis
Which ones are vector subspaces?
Example I.3. Let X₁, X₂, X₃ ⊂ R² be subsets defined as:
X₁ = {(x₁, x₂) ∈ R²: x₁ = x₂},
X₂ = {(x₁, x₂) ∈ R²: x₁= x₂ + 1},
X₃ = {(x₁, x₂)∈ R² : x₁=x₂²}
X1 is a vector subspace, but X2 and X3 are not.
1) check that if x = (x₁, x₂)
and y = (y₁, y₂) are vectors from X1, then any linear combination ax + by also lies in X1.
ax + by = a(x₁, x₂) + b(y₁, y₂)
= (ax₁+ by₁ , ax₂ + by₂)
Since x₁= x₂ and y₁=y₂ we conclude that ax₁ + by₁ = ax₂ + by₂ , that is ax + by lies in X1.
2) The set X2 is not a vector space, since it does not contain the zero vector 0 = (0, 0); by Definition I.2 any vector space should do so.
3) X3 is not a vector space, note that the vectors (1, 1) and (1, −1) lie in X3, but the sum (2, 0) = (1, 1) + (1, −1) does not.
orthogonal
when vectors are orthogonal angle is (pi/2)k
ie x.y=0
pythagorean thm
applies to sides which are orthogonal
IFF |x+y|^2 = |x|^2+|y|^2
def 1.4 linearly dependent
A collection of non-zero vectors x₁, . . . , xₖ ∈ Rⁿ is called linearly dependent if there
exists a collection of real numbers α₁, . . . , αₖ ∈ R such that not all of the αᵢ’s are equal to zero and
α₁x₁ + α₂x₂ + · · · + αₖxₖ = 0.
A collection of vectors that are not linearly dependent is called linearly independent.
Self-Check Question I.4. Let x₁, . . . , xₖ ∈ Rⁿ be a collection of non-zero vectors that are pair-wise orthogonal, that is
x_i·x_j = 0 for all i not equal to j.
Can you show that these vectors are linearly independent?
suppose linearly dependent
consider
a₁x₁+ . . . +aₖxₖ=0
we look for these to be non zero reals
Taking x_i * (a₁x₁+ . . . +aₖxₖ)=0
gives only non zero dot prod as
a_i (x_i x_i)=0
we have x_ix_i not equal to 0 as these are non zero vectors and |x_i|^2 = x_i*x_i
thus must have a_i=0
repeat argument for all a_i
thus linearly independent
x,y in R linearly dependent if
2 vectors linearly dependent IFF x=0 or y=kx k in R (covector)
scalar multiples a_1x+a_2y=0 s.t a_1 not equal to 0 x=(a_1/a_2)y=0
spans
Let x₁, . . . , xₖ ∈ Rⁿ be a collection of non-zero vectors
collection spans a vector subspace X ⊂ R^n if for any x ∈ X there exist k real numbers αi ∈ R, where i = 1, . . . , k, such that
x = α₁x₁ + α₂x₂ + · · · + αₖxₖ
Let x₁, . . . , xₖ ∈ Rⁿ be a collection of non-zero vectors that lie in a vector subspace X ⊂ R^n. Then EQUIVALENT STATEMENTS
(i) x1, . . . , xk is a maximal system of linearly independent vectors in X;
(ii) x1, . . . , xk is a minimal system of vectors that span X;
(iii) x1, . . . , xk is a system of linearly independent vectors that span X;
(iv) for any x ∈ X there exist a unique collection of real numbers α1, . . . , αk such that
x = α1x1 + α2x2 + · · · + αkxk.
def 1.5 BASIS
Let X ⊂ R^n be a vector subspace. A collection of non-zero vectors x1, . . . , xk ∈ X that satisfies one of the equivalent statements (i)-(iv) in Proposition I.3 is called a basis of X.
a basis exists of at most n vectors in R^n
we prefer an orthonormal basis
prefer an orthonormal basis
If its a basis then any vector x in V can be expressed as a unique linear combo
x= a_1x_1 +…+a_kx_k for a_i in R
if basis is orthonormal then coeffs are easier to compute
a_i = (x*x_i) bc dot prods =1 or =0 for that i’th component
DIMENSION of a vector space X
BASIS VECTORS
f x_1, . . . , x_k and y1, . . . , ym are two different basis of X, then k = m,
the number of vectors in a basis of a fixed vector space X is always the same.
This integer is the dimension of a vector space X.
E.G Let X,Y be 2 vector spaces in R^n
Show that if X⊆ Z and dim Z= dim X
then X=Z
The dimensions mean we can describe using a basis with the same number of vectors
Let x∈X then
x=a_1x_1+….+a_kx_k for some reals basis {x_1,…,x_k}
X⊆Z means x∈Z so
x=b_1z_1+….+b_kz_k for some reals basis {z_1,..z_k}
that means
the basis for Z is also a basis for X, as this is true for any x in X
take z in Z,
we use the basis
z=c_1z_1+….+c_kz_k
as this is a basis in X also, z in X
so Z⊆X
so X=Z
ONE DIMENSIONAL SUBSPACE
in R^n
ℓ_v={tv: t∈R }
v ∈R^n fixed non zero DIRECTION vector
Lines through O
basis 1 vector
Plane through 0∈R^n
vector subspace dim 2
vector subspace dim 2
Plane through 0∈R^n
ℓ_v ⊥
{(x,y,0,…,0} x,y ∈R^n}
L5: theory existence of orthonormal basis
DEFN ORTHOGONAL collection
Existence of an orthonormal basis. Let V⊂R^n be a subspace.
A collection of vectors x₁,…,xₖ ∈ Rⁿ is called ORTHOGONAL if vectors are pairwise orthogonal
xᵢ*xⱼ=0 for all i not equal to j
Note that if a collection of vectors is orthogonal, then the vectors are linearly independent.
ORTHONORMAL collection
A collection of vectors x₁,…,xₖ ∈ Rⁿ is called ORTHONORMAL if
xᵢ*xⱼ=
{1 if i=j
{0 if i not equal to j
PAIRWISE ORTHOGONAL
UNIT LENGTH
L5:
EXISTENCE OF A BASIS
Theorem I.3.
there MAYBE a question on the exam about this :)
Let V ⊂ Rⁿ be a vector subspace. Then there exists an orthonormal basis in V .
ORTHOGONAL COMPLEMENT OF X
X⊥
={y∈R^n y*x =0 ∀x∈X}⊂R^n
vector subspace X
all vectors orthogonal to all x∈X
dimensions of X and X⊥ in R^n
complimentary
dim X=k
dim X⊥=n-k
For ℓ_v through O in R^n
ℓ_v ⊥ is
a PLANE ORTHOGONAL to ℓ_v
is
X⊥ a VECTOR SUBSPACE?
yes
0 in X⊥
for x,u,v in X⊥
(au+bv)x= aux+bv*x=0 closed
(checking orthogonal to x in X checks its in our set )
V=span(e_1.e_2)
then V⊥
is a line formed by e_3 axis
orthogonal to e_1 and e_2
thus orthogonal to all combos of these (the plane)
(X⊥)⊥
=X
complementary dims
y in (X⊥)⊥ by definition is orthogonal to all vectors orthogonal to all in X
dim(X)=k
dim(X⊥)=n-k
thus
dim(X⊥)⊥ = n-(n-k)=k
x in X is not in X⊥ but is orthogonal to all y in X⊥
LEMMMA I.4
two lines coincide as sets IFF
Two lines
ℓ_{p,v}
ℓ_{q,w}
coincide as sets
ℓ_{p,v}=ℓ_{q,w}
IFF
the direction vectors v,w are linearly dependent and p-q=t_o v for some t_0 in R
PROOF
Two lines
ℓ_{p,v}
ℓ_{q,w}
coincide as sets
ℓ_{p,v}=ℓ_{q,w}
IFF
the direction vectors v,w are linearly dependent and p-q=t_o v for some t_0 in R
suppose lines coincide
q∈ℓ_{p,v}
hence there exists t_0∈R s.t
q= t_0 v +p
also as lines coincide
q+w∈ℓ_{p,v}
there exists t s.t
q+w=tv+p
ℓ_{q,w}hence
(t-t_0)v=w
thus v&w linearly dependent
conversely if
q=p+t_0v for some t_0 in R
then q∈ℓ_{p,v}
since the non zero vectors v and w are linearly dependent we can assume
w=av for non zero a
concluding
ℓ_{q,w}=
{q + sw : s ∈ R} =
{p + (t_0 + sα)v : s ∈ R} =
=ℓ_{p,v}
Proposition I.5.
The lines, viewed as subsets of R^2 of form
ℓ_{p,v}
={p+tv: t ∈ R} ⊂ R^2
satisfy Euclid’s axioms A1-A5.
PROOF
EUCLIDS AXIOMS
A1. For any two points there exists a line that goes through both of them.
A2. There exists at most one line that passes through two distinct points.
A3. Every line contains at least two distinct points.
A4. There exist three points that do not lie on a straight line.
A5. (Parallel axiom). Let ` be a line and p a point that does not lie on . Then there exists a unique
line that contains p and does not intersect the line
.
LINEAR MAP/OPERATOR
defn
A map L : V → W is called linear if
L(α_1u_1 + α_2u_2) =
α_1L(u_1) + α_2L(u_2)
for all α_1, α_2 ∈ R, u_1, u_2 ∈ V.