Week 2 Flashcards

1
Q

What is the multivariate MSE?

A

Var(theta^hat) + (bias(theta hat))(bias(theta hat)’

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the multivariate normal distribution?

A

f(x) = 2 pi ^(-n/2) |𝝨|^(1/2) exp(-(1/2) (x-mu)’ 𝝨^-1(x-mu))

With mu = vector of mu
And Sigma = diagonal matrix of variances

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How can a (multivariate) normal distribution be rewritten to a chi squared distribution?

A

(x - μ)’Σ^-1(x-μ) or

If x is N(0, I.n) distributed and matrix A is symmetric idempotent of rank r then

x’Ax

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How can a Normal and chi squared distribution be rewritten to a students t-distribution?

A

x/sqrt(w/n) is t(n) distributed, this converges to a normal distribution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How can two chi squared distributions be rewritten to a F-distribution?

A

(v/m)/(w/n) is F(m, n) distributed (given that v and w are independent.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How can a Student’s t-distribution be rewritten to a F-distribution?

A

t^2 = F(1, n) distribution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What can be said about the limit of a F-distribution?

A

If w/n -> inf then F(m, n) is (chi squared (m))/m distributed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the derivation of the Least Squares estimation?

A
  1. We need to choose a β s.t. the error is minimized, so we create function ɸ(β) = sum of ui2 = u’u = (y - Xβ)’(y-Xβ)
  2. We want to minimize this function, so first we rewrite it to:

ɸ(β) = y’y = y’Xβ - β’X’y + β’X’Xβ

= y’y - 2y’Xβ + β’X’Xβ (because y’Xβ = β’X’y)

  1. Then we differentiate and equate to 0:

derivative of ɸ(β) with respect to β = -2y’X + 2β’X’X = 0

gives

X’Xb = X’y, so b = (X’X)-1X’y

  1. So we get yhat = Xb = X(X’X)-1X’y
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the residual maker matrix, and why is it named like that?

A
  1. M = In - X(X’X)-1X’, this is a symmetric idempotent matrix (AA=A).
  2. It’s derived from the residuals. i.e. e = y - Xb = y - X(X’X)-1X’y = My
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the formula for b?

A

(X’X)-1X’y

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the least absolute deviations estimator?

A

Instead of minimizing the sum on ui2, we now minimize the sum of the absolute value of ui.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What are the advantages and disadvantages of the least absolute deviations?

A

Advantages:

  • Resistant to outliers

Disadvantages:

  • Might have multiple solutions
  • Not differentiable, so it’s a huge disadvantage in theoretical results
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How is R2 derived?

A
  1. We have that y = Xb + e = yhat + e
  2. Since X’e = 0, we have yhate = b’X’e = 0, thus y’y = yhatyhat + e’e. So we can create the ratio:

(yhatyhat)/(yhatyhat) = 1 - (e’e)/(yhatyhat)

  1. If we create the symmetric idempotent matrix A (with Ae = e), then

Ay = AXb + Ae = AXB + e, so then

y’Ay = b’X’AXb + e’e (a.k.a. SST (a.k.a. TSS) = SSE + SSR)

  1. Thus we can usually write R2 = 1 - SSR/SST (aka. TSS) = 1 - (e’e)/(y’Ay)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

(AB)’ = … ?

A

B’A’

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

(AB)-1 = … ?

A

B-1A-1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is ɩ?

A

It’s a vector of ones.