Reading Quiz 15 Flashcards
least-squares regression
fits a straight line to data in order to predict a response variable y from the explanatory variable x
inference about regression conditions
- the observations must be independent
- the true relationship is linear
- the standard deviation of the response about the true line is the same everywhere
- the response varies normally about the true regression line
the observations must be independent condition
in particular, repeated observations on the same individual are not allowed
the true relationship is linear condition
we can’t observe the true regression line, so we will almost never see a perfect straight-line relationship in our data
look at the scatter plot to check that the overall pattern is roughly linear
a residual plot against x magnifies any unusual pattern
the standard deviation of the response about the true line is the same everywhere condition
look at the residual plot
the scatter of the data points (the vertical distance) about the y=0 line should be roughly the same over the entire range of the data
the response varies normally about the true regression line condition
the residuals estimate the deviations of the response from the true regression line, so they should follow a normal distribution
make a boxplot, histogram, or stemplot of the residuals and check for clear skewness or other major departures from normality
slight departures from normality do not greatly affect inference for regression, so they are allowed, particularly when we have many observations
regression model
says that there is a true regression line μy = α + βx that describes how the mean response μy varies as x changes
true regression line
μy = α + βx
describes how the mean response μy varies as x changes
the observed response y for any fixed x has a normal
σ for any value of x
the parameters of the regression model are
the intercept α estimated by a, the slope β estimated by b, and the standard deviation σ estimated by s
the true slope β says how much
change in y when x increases by 1
the standard deviation σ describes
how much variation there is in responses y when x is fixed
to estimate σ
use the standard error about the line, s
s
regression standard error
s= sqrt((Σresiduals^2)/(n-2)) = sqrt((Σ(y-yhat)^2)/(n-2))
sample standard deviation of the residuals
spread of data (measure of variability) around the least squares regression line
“typical” amount of prediction error when using a linear regression model to make predictions
calculator compute s
enter data into L1 and L2
STATS, TESTS, LinRegTTest
regression standard error has how many degrees of freedom
n - 2
all t procedures in regression inference have n-2 degrees of freedom
inference for regression goal
predict behavior of y for given values of x
inference for regression cont
there is an “on average” straight line relationship between y and x
saying μy moves along a straight line as explanatory variable x changes