Statistical Models Flashcards
Define parameters in regression equations (slope, Y-intercept)
The intercept and weight (measurement error) values are called the parameters of the model.
What are the weight values?
Regression weights, or regression coefficients. Weight values are values that assume for the amount of error/random error in the model. They are calculated using the slope of the line, where the dependent Y [reliant] has a quantitative relationship with the independent X [controlled]. Y is assumed to have a constant standard deviation over multiple observations, which can then minimize error by calculating least squared estimates [or taking correlation by SD of Y compared to SD of X].
What is the intercept?
Used to account for error when the mean of the residuals [Y] is theoretically equal to zero [X=0]. It tells you nothing about the relationship b/w X and Y, and serves as a constant that gives some idea of where Y crosses the y-axis.
What do these two parameters estimate?
The slope and the intercept define the linear relationship between two variables, and can be used to estimate an average rate of change.
What are predicted values?
The values of Y-hat, or the values of the dependent variable assumed by the parameters of the model.
What are observed values?
The values of Y, or the actual values of the dependent variable.
How should we understand model fit?
Model fit is a measurement/representation of how well the actual values of Y correspond to the predicted values of Y.
What specific equation is used to assess/estimate model fit?
Error variance, or the average squared deviations of the model parameters. This is used because the degree of error in the model tells us how far the predicted values distance from reality.
What is good error variance?
Small, minor differences in variability that do not detract or misrepresent the relationship b/w Y and Y-hat. Good error variance is not necessarily explainable, but also does not impact how accurate the model is assumed to be or can be corrected for.
What is bad error variance?
Differences in variability that are not explainable and that confound the relationship b/w Y and Y-hat. In this case, the error that we are attempting to examine can be related to entirely different variables, or are not actually relevant to the question that the model is attempting to represent.
What does least squares estimation do?
Finds a line determined by using an estimation of squared SD error variance relative to the constant Y [if the mean of X=0]. It makes the sum of squared errors as small as possible, and thus minimizes the total compounded error.
What is a residual?
Error estimates that result from the model being unable to perfectly reconstruct the actual data b/c it cannot realistically represent the full population error.
What is error variance?
The average squared differences b/w the predict and the observed values.
What is R^2?
A squared correlation that represents the proportion of the variance in Y that is accounted for by the model. It estimates the strength of the relationship b/w the model and the response variable.
Explain the purpose of statistical modeling.
To provide a mathematical representation of theories [about the relationships b/w different factors].