lecture 1 - matrix algebra Flashcards

1
Q

properties of matrices

A
  1. inverse property
  2. transpose of product
  3. inverse of product
  4. transpose of inverse
  5. double transpose
  6. distributive property for transpose
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is a Symmetric Matrix?

A
  • satisfies A^⊤ = A
  • means the matrix is equal to its transpose
  • Elements across the main diagonal are mirrored.
  • mathematically: a_{ij} = a_{ji}
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Definition of Positive Definite Matrix

A

a matrix A is positive definite if x^⊤Ax > 0 for all non-zero vectors x, given that A is symmetric

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Why are Positive Definite Matrices Important in Machine Learning?

A
  • Indicates strictly convex functions, leading to optimal solutions.
  • In machine learning, this guarantees that algorithms like linear regression converge to a single, optimal set of parameters.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the result of multiplying a matrix and a vector?

A
  • When you multiply a matrix A and a column vector x, the result is a new column vector.
  • Each element of the new vector corresponds to a weighted sum of the matrix’s rows and the vector’s elements.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How do you compute the elements of the result when multiplying a matrix by a vector? (Ax_i)

A

To find the value of a specific row in the resulting vector:

  1. Multiply each element in that row of the matrix by the corresponding element in the column vector.
  2. Add these products together to get the final value.
  • (Ax)i = \sum^n{j=1} a_{ij}x_j
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What does each element in the resulting vector Ax represent?

A
  • Each element represents the “weighted sum” of one row of the matrix, weighted by the values of the column vector.
  • row of matrix * vector column
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What does differentiating a scalar with respect to a vector produce?

A
  • Produces another vector.
  • The resulting vector contains partial derivatives of the scalar with respect to each element of the original vector.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How do you differentiate a scalar-vector product with respect to a vector?

A
  • e.g., x^T a
  • only the terns involving the current varianble x_i survive
  • the result is a transposed vector (i.e., a row vector)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Why do we transpose the result of vector differentiation?

A

The transpose is necessary because gradients are conventionally row vectors. This ensures that when multiplied by another vector, the matrix dimensions align correctly.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What happens when you differentiate b^T Ax with respect to x?

A
  • A^T b
  • This result comes from applying the scalar-vector differentiation rule, where only terms involving the current variable x_i survive, and the coefficients form the product A^T b
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What does differentiating a vector with respect to a vector produce?

A

A jacobian matrix

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the Jacobian matrix?

A
  • The Jacobian matrix is the derivative of a vector-valued function with respect to a vector of variables.
  • It describes how each component of the function changes with respect to each variable.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What does the Hessian matrix represent?

A
  • The Hessian matrix is a square matrix of second-order partial derivatives of a scalar-valued function.
  • It provides information about the curvature of the function.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How does the Hessian matrix help determine the nature of a critical point?

A
  • Positive eigenvalues: The function is convex, indicating a local minimum.
  • Negative eigenvalues: The function is concave, indicating a local maximum.
  • Mixed eigenvalues: The function has a saddle point, meaning it curves up in one direction and down in another.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the derivative of a vector with respect to itself?

A
  • I
  • The identity matrix
17
Q

What is the derivative of Ax with respect to x?

A

A

18
Q

What is the derivative of x^T A with respect to x?

A

A^T

19
Q

What is the derivative of x^T Ax with respect to x?

A
  • x^T(A+A^T) if A is not symmetric
  • 2x^T A if a is symmetric
20
Q

What is the second derivative of x^T Ax with respect to x?

A
  • A+A^T if A is not symmetric
  • 2A if a is symmetric
21
Q

proof for the derivative of a vector with respect to itself being I

A

Each element (i,j) of the result is:

  • 1 if i = j
  • 0 if i =/= j
  • Thus, the result is the identity matrix.