Ch 7 Notes Flashcards

1
Q

7.1 Linear Regression Model

Captures relationship between two or more variables

Predicts an outcome of a target variable based on several input variables

Make assessments and robust predictions by determining which of the relationships matter the most and which we can ignore

A

Regression Analysis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

7.1 Linear Regression Model

Formulate a mathematical model that relates the outcome of a target variable, called the response variable, to one or more other input values called predictor values

Use information on the predictor variables to describe and/ or predict changes in the response variable

A

Predictor Variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

7.1 Linear Regression Model

Known to perform well for making predictions

Fail to establish a cause-and-effect relationship between the variables because of the nonexperimental nature of most applications

May appear to search for casuality when it basically detects correlation

Casuality: can only be established through randomized experiments and/ or advanced statistical models

A

Regression Model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

7.1 Linear Regression Model

Cannot expect to predict its exact (unique) value

If value is uniquely determined by the values of the predictor variables, the relationship between the variables is deterministic

Relationship between this variable and predictor variable is stochastic, due to the omission of relevant factors that influence this variable

A

Response Variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

7.1 Linear Regression Model

How to develop a mathematical model that captures the relationship between the response variable 𝑦 and the π‘˜ predictor variables π‘₯_1,π‘₯_2,β‹―,π‘₯_π‘˜.

*Must account for randomness that is a part of real life

A
  1. Start with a deterministic component that approximates the relationship
  2. Then add a random error term to it, making the relationship stochastic
  3. Use economic theory, intuition, and statistical measures to determine which predictor variables might best explain the response variable
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

7.1 Linear Regression Model

Assumes meaningful numeric values

A

Numerical Value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

7.1 Linear Regression Model

Reflect categories

A

Categorical Variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

7.1 Linear Regression Model

Used to describe two categories of a categorical variable, denoted d

Indicator or binary variable
𝑑=1 for one of the categories
𝑑=0 for the other(s)
The category with 𝑑=0 is called the reference or benchmark category
All comparisons are made relative to this category

A

Dummy Variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

7.1 Linear Regression Model

**Uses only 1 predictor value

𝑦=𝛽_0+𝛽_1 π‘₯_1+πœ€

𝛽_0 is the unknown intercept

𝛽_1 is the unknown slope

𝛽_0+𝛽_1 π‘₯_1 is the deterministic component

πœ€ is the stochastic component or random error term

**The expected value of 𝑦 for a given value of π‘₯_1 lies on a straight line: 𝐸(𝑦)=𝛽_0+𝛽_1 π‘₯_1.

A

Simple Linear Regression Model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

7.1 Linear Regression Model

𝛽_1 > 0

A

Positive Linear Relationship

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

7.1 Linear Regression Model

𝛽_1< 0

A

Negative Linear Relationship

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

7.1 Linear Regression Model

𝛽_1 = 0

A

No linear Relatioship

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

7.1 Linear Regression Model

Model with more than one predictor

𝑦=𝛽_0+𝛽_1 π‘₯_1+𝛽_2 π‘₯_2+…+𝛽_π‘˜ π‘₯_π‘˜+πœ€

Replace π‘₯ with 𝑑 for a dummy variable.

The population parameters 𝛽_0,𝛽_1,𝛽_2,…,𝛽_π‘˜ are unknown and must be estimated using sample data.

Sample data: 𝑛 observations of 𝑦,π‘₯_1,π‘₯_2,…,π‘₯_π‘˜

Use the sample data to obtain 𝑏_0,𝑏_1,𝑏_2,…𝑏_π‘˜ which are estimates of 𝛽_0,𝛽_1,𝛽_2,…,𝛽_π‘˜.

A

Multiple Linear Regression Model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

7.1 Linear Regression Model

Predicted value of the response variable given specified values of predictor varibles

A

𝑦̂ (read as y-hat)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

7.1 Linear Regression Model

𝑒=π‘¦βˆ’π‘¦Μ‚

Difference between observed and predicted values

A

Residual

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

7.1 Linear Regression Model

𝑆𝑆𝐸=βˆ‘(π‘¦βˆ’π‘¦Μ‚ )^2 = βˆ‘π‘’^2

Chooses the sample regression equation by minimizing the error sum of squares

Desirable properties is certain assumptions hold

Gives an equation β€œclosest” to the data

A

Ordinary Least Squares (OLS)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

7.1 Linear Regression Model

𝑏_0 is the estimate of 𝛽_0

𝑏_𝑗 is the estimate of 𝛽_𝑗

A

Estimated regression coefficients

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

7.1 Linear Regression Model

Predicted value of 𝑦̂ when each predictor variable assumes a value of 0

Not always meaningful

A

𝑏_0 is the estimate of 𝛽_0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

7.1 Linear Regression Model

Change in the predicted value of the response given a unit increase in π‘₯_𝑗, holding all other predictor variables constant

Partial influence of π‘₯_𝑗 on 𝑦̂

A

𝑏_𝑗 is the estimate of 𝛽_𝑗

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

7.1 Linear Regression Model

Subject to sampling variability

Will change if we use a different sample to estimate the regression model

Always wider than the corrisponding confidence interval

A

Predictions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

7.1 Linear Regression Model

point estimate Β± margin of error

Use a confidence interval as the interval estimate for the mean (expected value) of y

Use a prediction interval as the interval estimate for the individual values of y

We use the same point estimate for constructing both

A

Interval Estimate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

7.1 Linear Regression Model

𝑦̂^0±𝑑_(π›Όβˆ•2,𝑑𝑓) 𝑠𝑒(𝑦̂^0)

𝑦̂^0=𝑏_0+𝑏_1 π‘₯_1^0+ 𝑏_2 π‘₯_2^0+ β‹― + 𝑏_π‘˜ π‘₯_π‘˜^0

𝑑𝑓=π‘›βˆ’π‘˜βˆ’1

One way to obtain this model is to estimate a modified regression model

y is the response variable

Explanatory variables defined as π‘₯_𝑗^βˆ—=π‘₯_π‘—βˆ’π‘₯_𝑗^0

The resulting estimate of the intercept and its standard error are 𝑦̂^0 and 𝑠𝑒(𝑦̂^0)

A

For specific values π‘₯_1^0,π‘₯_2^0,β‹―,π‘₯_π‘˜^0, the 100(1βˆ’π›Ό)% confidence interval for the Expected value of y

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

7.1 Linear Regression Model

𝑦̂^0±𝑑_(π›Όβˆ•2,𝑑𝑓) √(γ€–(𝑠𝑒(𝑦̂^0))γ€—^2+𝑠_𝑒^2 )
𝑦̂^0=𝑏_0+𝑏_1 π‘₯_1^0+ 𝑏_2 π‘₯_2^0+ β‹― + 𝑏_π‘˜ π‘₯_π‘˜^0

𝑑𝑓=π‘›βˆ’π‘˜βˆ’1

Prediction Interval is wider than the confidence interval

Prediction Interval Incorporates the variability of the random error term

Higher variability makes it more difficult to predict accurrately, necessitating a wider interval

A

For specific values π‘₯_1^0,π‘₯_2^0,β‹―,π‘₯_π‘˜^0, the 100(1βˆ’π›Ό)% prediction interval for the Individual Value of y

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

7.1 Linear Regression Model

*A categorical variable may be defined by more than two categories

*Use multiple dummy variables to capture all categories, one for each category

*Given the intercept term, we exclude one of the dummy variables from the regression

-Excluded variable represents the reference category

-Including all dummy variables creates perfect multicollinearity

A
25
Q

7.2 Model Selection

Summarize how well the sample regression equation fits the data

Standard error of the estimate: 𝑠_𝑒

The coefficient of determination, 𝑅^2

The adjusted coefficient of determination, adjusted 𝑅^2

A

Goodness of Fit Measures

26
Q

7.2 Model Selection

Difference between the observed and predicted value of the response, 𝑒_𝑖=𝑦_π‘–βˆ’π‘¦Μ‚_𝑖

Sample regression equation provides a good fit when the dispersion of residuals is relatively small

Sample variance: 𝑠_𝑒^2 is the average squared deviation between the observed and predicted values

A

Residuals

27
Q

7.2 Model Selection

Has the same units of measurement as the response

𝑠_𝑒=√(𝑆𝑆𝐸/(π‘›βˆ’π‘˜βˆ’1))

𝑆𝑆𝐸 is the error sum of squares, π‘˜ denotes the number of predictors, 𝑛 is the sample size.

A

Standard Deviation of Residuals (standard error of the estimate)

28
Q

7.2 Model Selection

For a given sample size, increasing the number of predictors reduces the numerator and denominator

The net effect allows us to determine if the added predictor variables improve the fit

When comparing models with the same response, the model with the smaller 𝒔_𝒆 is preferred

A
29
Q

7.2 Model Selection

Quantifies the sample variation in the response that is explained by the sample regression equation

Ratio of the explained variation of the response variable to its total variation

A

Coefficiant of Determination 𝑅^2

30
Q

7.2 Model Selection

Use in the context of regression

The total variation in 𝑦 is the total sum of squares, 𝑆𝑆𝑇=βˆ‘(𝑦_π‘–βˆ’π‘¦Μ… )^2

Break into two parts: explained variation and unexplained variation

𝑆𝑆𝑇=𝑆𝑆𝑅+𝑆𝑆𝐸

The variation in 𝑦 explained by the sample regression equation is the regression sum of squares,
𝑆𝑆𝑅=βˆ‘(𝑦̂_π‘–βˆ’π‘¦Μ… )^2

A

Analysis of Variance (ANOVA)

31
Q

7.2 Model Selection

The proportion of the sample variation in the response explained by the sample regression equation

Falls between 0 and 1

The closer to 1, the better the fit

A

𝑅^2=𝑆𝑆𝑅/𝑆𝑆𝑇=1βˆ’π‘†π‘†πΈ/𝑆𝑆𝑇

32
Q

7.2 Model Selection

Can’t use 𝑅^2 when the competing models do not include the same number of predictor variables (but have the same response)

𝑅^2 never decreases as we add more variables

May include variables with no economic or intuitive foundation

A

Comparing Models

33
Q

7.2 Model Selection

Explicity accounts for the sample size n and the number of predictor variables k

1βˆ’(1βˆ’π‘…^2 )((π‘›βˆ’1)/(π‘›βˆ’π‘˜βˆ’1))

Imposes a penalty for any additonal predictors

The higher the adjusted 𝑅^2, the better the model

When comparing models with the same response, the model with the higher is preferred

A

Adjusted 𝑅^2

34
Q

7.2 Model Selection

𝑦=𝛽_0+𝛽_1 π‘₯_1+𝛽_2 π‘₯_2+…+𝛽_π‘˜ π‘₯_π‘˜+πœ€

*We can conduct hypothesis tests about the unknown parameters 𝛽_0,𝛽_1,𝛽_2,…,𝛽_π‘˜.

-Joint test about all of the parameters
-Individual tests about a single parameter

*For the tests to be valid, 𝑏_0,𝑏_1,𝑏_2,…𝑏_π‘˜ must be normally distributed.

-Satisfied if the random error term πœ€ is normally distributed

-If πœ€ is not normally distributed, the tests are valid only for large sample sizes.

A

Linear Regression Model with k predictor variables

35
Q

7.2 Model Selection

Regarded as a test of the overall usefulness of a regression

Determines whether the predictor variables have a joint statistical influence on the responds

𝑦=𝛽_0+𝛽_1 π‘₯_1+𝛽_2 π‘₯_2+…+𝛽_π‘˜ π‘₯_π‘˜+πœ€

-All of the slope coefficients equal zero; all of the predictors drop out; none of the predictors have a linear relationship with the response

-At least one of the slope coefficients does not equal zero; at least one predictor influences the response

𝐻_0:𝛽_1=𝛽_2=…=𝛽_π‘˜=0
𝐻_𝐴:𝐴𝑑 π‘™π‘’π‘Žπ‘ π‘‘ π‘œπ‘›π‘’ 𝛽_𝑗≠0

Test statistic measures how well the sample regression equation explains the variability in the response

A

Test of Joint Significance

36
Q

7.2 Model Selection

𝐹_(𝑑𝑓_1, 𝑑𝑓_2 ) =(π‘†π‘†π‘…β„π‘˜)/(𝑆𝑆𝐸⁄(π‘›βˆ’π‘˜βˆ’1)=𝑀𝑆𝑅/𝑀𝑆𝐸

𝑑𝑓_1=π‘˜
𝑑𝑓_2=π‘›βˆ’π‘˜βˆ’1
𝑆𝑆𝑅 is the regression sum of squares.
𝑆𝑆𝐸 is the error sum of squares.
𝑀𝑆𝑅 is the mean square regression.
𝑀𝑆𝐸 is the mean square error.

A

Test Statistic

37
Q

7.2 Model Selection

p-value: 𝑃(𝐹_(𝑑𝑓_1, 𝑑𝑓_2 ) β‰₯𝑀𝑆𝑅/𝑀𝑆𝐸)

-A larger test statistic provides us with more evidence to reject the null hypothesis

-A larger test statistic indicates that a large portion of the sample variation in the response is explained by the regression model, thus the model is useful

A

Right Tailed F Test

38
Q

7.2 Model Selection

*If 𝛽_1=0, then π‘₯_1 basically drops out of the equation.

There is no linear relationship between π‘₯_1 and 𝑦.

π‘₯_1 does not influence 𝑦.

A

Linear regression model with π‘˜ predictor variables

𝑦=𝛽_0+𝛽_1 π‘₯_1+𝛽_2 π‘₯_2+…+𝛽_π‘˜ π‘₯_π‘˜+πœ€

39
Q

7.2 Model Selection

-Two-tailed
-Right-tailed
-Left-tailed

Typically 𝛽_𝑗0=0, but it could be a nonzero value.

A

Let 𝛽_𝑗0 be a hypothesized value of 𝛽_𝑗

40
Q

7.2 Model Selection

H_0: B_j = B_j0
H_A: B_j does not equal B_j0

A

Two Tailed Test

41
Q

7.2 Model Selection

H_0: B_j less than or equal to B_j0

H_A: B_j greater than B_j0

A

Right Tailed Test

42
Q

7.2 Model Selection

H_0: B_j greater than or equal to B_j0

H_A: B_j less than B_j0

A

Left Tail Test

43
Q

7.2 Model Selection

𝑑_𝑑𝑓=𝑏_π‘—βˆ’π›½_𝑗0/𝑠𝑒(𝑏_𝑗)

𝑑𝑓=π‘›βˆ’π‘˜βˆ’1

𝑏_𝑗 is the estimate for 𝛽_𝑗

𝑠𝑒(𝑏_𝑗) is the standard error of the estimator 𝑏_𝑗

𝛽_𝑗0 is the hypothesized value of 𝛽_𝑗

If the hypothesized value is zero,
𝑑_𝑑𝑓=𝑏_𝑗/𝑠𝑒(𝑏_𝑗)

Note that the testing framework for the intercept 𝛽_0 is similar.

Statistical packages report test statistics and p-values for a two-tailed test that assesses whether the coefficient differs from zero.

A

Test statistic for a test of individual significance

44
Q

7.2 Model Selection

Expresses the risk-adjusted return of an asset as a function of the risk-adjusted market return.

π‘…βˆ’π‘…_𝑓=𝛼+𝛽(𝑅_π‘€βˆ’π‘…_𝑓 )+πœ€

𝑅 is the rate of return on a stock or portfolio.

𝑅_𝑀 is the market return (typically the S&P 500).

𝑅_𝑓 is the risk-free interest rate (typically a Treasury bill).

𝛼 and 𝛽 are used in place of 𝛽_0 and 𝛽_1

A

Capital Asset Pricing Model (CAPM)

45
Q

7.2 Model Selection

Called the stock’s beta

𝛽=1: any change in the market leads to an identical change in the stock

𝛽>1: stock is more aggressive or riskier than the market

𝛽<1: stock is conservative or less risky than the market

A

𝛽: measures how sensitive the stock’s return is to changes in the market

46
Q

7.2 Model Selection

Called the stock’s alpha

𝛼>0: positive abnormal returns

𝛼<0: negative abnormal returns

A

𝛼: predicted to be zero, thus nonzero values indicate abnormal returns

47
Q

7.2 Model Selection

*Regression results are often reported in a table

A
48
Q

7.3 Model Assumptions and Common Violations

Statistical properties and validity of testing procedures depend on assumptions.

Under the assumptions, have desirable properties.

They are unbiased, 𝐸(𝑏_𝑗 )=𝛽_𝑗.

They have minimum variations between samples.

These are compromised when one or more of the model assumptions are violated.

*The validity of the significance test is also impacted

-Estimated standard errors are inappropriate
-Not possible to make meaningful inferences from the t and F tests

A

OLS Estimators

49
Q

7.3 Model Assumptions and Common Violations

Based on the error term πœ€

The residuals, or the observed error term, 𝑒=π‘¦βˆ’π‘¦Μ‚, contain useful information about πœ€

Use the residuals to investigate the through residual plots.

Residuals vs. predictor variable: nonlinear patterns

Residuals vs. predicted values 𝑦̂: changing variability

Residuals sequentially: correlated observations

Residuals can also be used to detect outliers.

A

Assumptions

50
Q

7.3 Model Assumptions and Common Violations

Observations that stand out from the rest of the data

Observation stands out in the residual plot

Can greatly impact the estimates

It is not always clear what to do with them

May indicate bad data due to incorrectly recorded observations

-These observations should be corrected or deleted

May just be due to random variations in which case the relevant observations should remain

A

Outliers

51
Q

7.3 Model Assumptions and Common Violations

*Plot the residuals on the vertical axis and the predictor variable or predicted values on the horizontal line

A
52
Q

7.3 Model Assumptions and Common Violations

Linearity is in the parameters not the in the variables

Make economic and intuitive sense

Include all relevant predictor variables

Incorporate nonlinearalities

Detection:
-Use residual plots to identify nonlinear patterns
-Linearity is justified if the residuals are randomly dispersed
-A discernible trend in the residuals indicates a nonlinear pattern

Remedy:
-Employ nonlinear regression
-Transform the response and predictor variables

A

Assumption 1: Regression Model is linear in the parameters and is correctly specified

53
Q

7.3 Model Assumptions and Common Violations

Exists when two or more predictor variables have an exact linear relationship

Difficult to disentangle the seperate influences of the predictor variables

Results in imprecise estimates of slope coefficients

Detection:
-Model can’t be estimated
-High 𝑅^2 and significant F statistic coupled with individually insignificant predictor variables
-High correlations between predictor variables
-Wrong signs of the estimated coefficients
-Insignificant important predictor variables

Remedy
-Drop one of the collinear variables
-Obtain more data
-Express variables differently
-Do nothing

A

Assumption 2: There is no exact linear relationship among the predictor variables, there is no multicollinearity

54
Q

7.3 Model Assumptions and Common Violations

-Constant variability

-No heteroskedasticity

-Often breaks down cross sectional data

-Exists when the variability in the response changes with at least one predictor value

-OLS estimators are still unbiased
-Standard errors are not appropriate
-Can’t use t and f tests

Detection
-Plot the residuals against each predictor variable or predicted values
-Residuals should be randomly dispersed
-Violated when the variability of residuals increases or decreases

Remedy
-Correct the standard errors
-Called White’s robust standard errors
-Not available in Excel
-R has packages

A

Assumption 3: Conditional on the values of the predictors, the variance of the error term is the same for all observations

55
Q

7.3 Model Assumptions and Common Violations

-No serial correlation

-Often breaks down with time series data

-OLS estimates are still unbiased
-Standard errors are innapropriate, distorted downward making the model look better than it really is

-The t and F tests may suggest the predictor variables are significant when they are not

Detection:
-Plot the residuals sequentially over time
-Should show no pattern
-Postive (or negative) residuals followed by postive (or negative) residuals consecutively
-Switching between positive and negative residuals

Remedy:
-Use Newey West robust standard errors in R

A

Assumption 4: Conditonal on the predictor variables, the error term is uncorrelated across observations

56
Q

7.3 Model Assumptions and Common Violations

Allows construction of interval estimates and hypothesis tests

If the error term is not normally distributed, then the interval estimates and hypothesis tests are valid only for large samples

A

Assumption 6: The error term is normally distributed

57
Q

7.3 Model Assumptions and Common Violations

Use to obtain a residual plot

Open the stores data file

Data > Data Analysis > Regression

For Input Y Range, select the sales observations and for Input X Range, select the Sqft observations. Select residual plots

Obtain a residual plot sequentially over time

Open Sushi data file

Data > Data Analysis > Regression

Given the regression output, select the residuals and choose Insert > Scatter

A

Exel

58
Q

7.3 Model Assumptions and Common Violations

Use to obtain White’s robust standard errors

Import the stores data file in a data frame (table) and label it myData

Use to obtain Newey West Robust Standard Errors

Import Sushi Data File

A

R