Orthogonality and Least Squares Flashcards
What does it mean for two vectors to be orthogonal?
v dot w = 0
What does it mean for a vector to be orthonormal?
Vectors are orthonormal if they are all unit vectors and are orthogonal to one another
Ex. u_i dot u_k = 0
but u_i dot u_i = 1
Provide a formula for the orthogonal projection of a vector onto V in a subspace of R^n.
proj(x) = x’’ = (u1 dot x)u1 + … (um dot x)um
where the vectors um form an ORTHONORMAL basis
Define the orthogonal complement of a subspace.
Consider a subspace V of R^n. The orthogonal complement V(perp) of V is the set of those vectors x in R^n that are orthogonal to all vectors in V. Note that V(perp) is the kernel of the orthogonal projection onto V
Shown: T(x) = proj(x) = x’’ = 0 meaning x = x(perp)
4 properties of orthogonal complements
1) the orthogonal complement V(perp) of V is a subspace of R^n
2) The intersection V and V(perp) consists of the zero vector
3) dim(V) + dim(V(perp)) = n
4) (V(perp))perp = V
How do you find the angle between two vectors?
angle = arccos((x dot y)/(length of x times length of y))
Prove that the length of the orthogonal projection of x is less than or equal to the length of the vector x.
Note x = x’’ + x(perp)
Now we want to prove llproj(x)ll <= llxll
llxll^2 = llproj(x)ll^2 + llx(perp)ll^2
Therefore it must be less than or equal since there is another term that must be added to proj(x) to be equal to x
What does the Cauchy-Schwartz inequality state?
lx dot yl <= llxll llyll
DO THE PROOF!
Describe the Gram-Schmidt process.
We want to construct an orthonormal basis from a given subspace of V of R^n. As defined before, an orthornormal basis consists of the unit vectors that are orthogonal to one another. To find the first unit vector we would divide its respective nonzero vector by its length. After the first vector is computed, each proceeding unit vector will be computed similarly, however, now we will use each vector's v(perp) to find the proceeding unit vector. Note v(perp) = v - proj(v) = v - (u1 dot v)u1
Provide an alternative characterization of orthogonal projections applying the concepts of least squares solutions
Consider a vector x in R^n and a subspace V of R^n. Then the orthogonal projection is the vector in V CLOSEST to x, in that
llx - proj(x)ll < llx-vll
for all v in V different from proj(x)
Formally, describe the least squares solution for an inconsistent linear system. What if the system is consistent?
Consider an inconsistent linear system Ax = b where A is an n x m matrix. The vector x* is called a least squares solution of this system is
ll b - Ax* ll <= ll b - Ax ll for all x in R^m.
If the system is consistent, then the least squares solutions are its exact solutions: The error llb-Axll is zero
What is the normal equation? How is it derived?
A^T(Ax) = A^T(b) is the normal equation of Ax = b ll b - Ax* ll <= ll b - Ax ll Ax* = proj(b) b - Ax* A^T(b - Ax*) = 0 A^T(b) = A^T(Ax*)
If A is an n x n matrix of the linear system Ax = b and ker(A) = {0}, what can you say about its least squares solution?
x* = (A^TA)^-1(A^T(b))