L3 - Least Squares Method Flashcards

1
Q

Give the Simple Linear Regression function…

A

Yb(x) = b0 + b1x + error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the formula for the Y-intercept of the regression line?

A

b0 = y median - (b1 * x median)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the formula for the slope of the regression line?

A

b1 = r * (sy / sx)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the goal of the simple regression function?

A

To achieve as small error as possible.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the formula for the error of the simple linear regression function?

A

Transpose the simple linear regression function with respect to E

Error = Yb(x) - b0 - b1x

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What does MSE find?

A

The average error between the regression line and the data points.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Why do we square the errors for MSE?

A
  • We only want to deal with positive values.
  • If we summed, the positive and negative values would cancel each other out.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the advantage of MSE?

A
  • Sensitive to anomalies, thus can reduce their effect.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the goal of MSE? What must we tune to obtain this?

A
  • Achieve an error as close to 0 as possible.
  • We tune b0 and b1 to do this.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the Least Square Method?

A
  • A method of linear regression in which we tune b0 and b1 in order to obtain the lowest SSE ( Sum of Squared Error )
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are the 4 steps to the Least Squares Method?

A
  1. Transpose the system of equations with respect to Error.
  2. Square and sum equations. Minimise as much as possible.
  3. Calculate partial derivatives with respect to b0 and b1
  4. Use elimination to solve the remaining system of equations.
  5. Result is b0 and b1 with the smallest error.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What do we compare the Least Squares Method against?

A

The mean of Y, which is the most basic regression line.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What are the 2 types of variations encountered in LSM?

A

Unexplained Variations - The error that can’t be explained by the independent variable.

Explained Variations - The error that can be explained by the independent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How do we calculate the Total Variation of LSM?

A

Total Variation = Unexplained variation + explained variation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the Sum of Squared Residual (SSR)?

A

Measures the level of variance in the error of a regression model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the Sum of Squared Error (SSE)?

A

Measures the difference between the predicted values and the error mean (MSE).

17
Q

What is the coefficient of determination?

A

Value between 0 and 1 that informs us of the prediction quality of the model.

1 is better.

18
Q

What is the formula for R^2?

A

R^2 = 1 - (SSE/TSS)

19
Q

In regression formulas, what is omega? When do we care about it?

A
  • Omega is the regression coefficients.
  • We care when we are establishing feature importance (interpreting model over prediction).
20
Q

What is the difference between Prediction and Interpretation?

A

Prediction: Using model to predict a value. We focus on performance metrics, not omega. Be careful about model becoming black box.

Interpretation: Using model to gain insight into data. We focus on parameters toe stablish feature impact.