Session 6 - Model Assessment Flashcards
What do we observe in statistical modelling?
How can this relationship be written?
A response variable Y and one or more different predictors X1, X2, …Xp.
and assume there is some relationship between Y and X1, X2, …Xp
𝒀=𝒇(𝑿)+𝜺
with
𝑋=”(X1, X2, …Xp)”
ε “= random error term which is independent of X and has a mean of 0”
𝑓(𝑋) “is a function and describes the systematic part between Y and X”
What does statistical learning refer to?
A set of approaches for estimating f()
In statistical learning f() can be what?
Unknown or known (or better “assumed”)
In statistical learning if f() is known, then we just need to estimate what?
The parameters:
Example: Linear regression
y = B0 + B1 x1 + B2 x2 + … + Bk xk + E
In statistical learning, what do we need to identify if f() is unknown?
Variables, linear and non-linear relationships and the best methodology (“machine”) to estimate f()
Examples: (feature) variable selection, polynomial regression, generalized additive models, regularized regression, random forests, support vector machines
If our aim is inference what are we describing?
This is usually summarised as what?
The way Y is affected by changes in X
Usually summarized as average changes (If we increase x1 by 1 unit, Y will change on average by b1 units.
What is the aim of inference?
Understanding how Y changes as function of X and we want to know the exact form of f()
What kind of estimator do we want from an inferential model?
One that does not differ from the population parameter in a systematic manner (unbiased)
Consistent unbiased estimator reach the true with increasing sample size
Among unbiased estimator we choose the one with the minimum sampling variance (efficient).
If we estimate the parameter many times, than the mean of all parameters is close to true population mean and the variance of all estimates is the smallest possible!
Unbiased estimator can be mean or median but has been shown that mean has less sampling variance than median
What does the Gauss-Markhov Theorem tell us?
Within the class oflinear and unbiasedestimators, the Ordinary Least Square (OLS) estimator is most efficient.
How is the Best linear unbiased estimator (BLUE) estimated?
By minimizing the sum of squares of the differences between observed and predicted values in a given data set (OLS method).
With a given data set why does the OLS provide the smallest confidence intervals of all unbiased estimators?
Because it is unbiased, it has the smallest possible Mean Squared Error (MSE)within the linear and unbiased class of estimators.
How can we estimate prediction error?
An estimate of prediction accuracy for continuous outcomes is the Mean Squared Error (MSE):
𝑀𝑆𝐸= (∑((𝑌𝑖)̂− 𝑌𝑖)2)/𝑛 with
Y^𝑖= estimated Y for case i
Yi = observed Y for case i
N = number of cases in hold out sample
What are the assumptions for a linear regression?
Linearity: the response variable Y is linearly related to the independent variables (X’s)
Independence: errors (and hence the observations on Y) are independent of each other
Normality: errors (and hence the observations on Y) are normally distributed
Homoscedasticity: errors (and hence Y’s) have constant variance
What is important for validity of inference?
Assumptions
What happens if the assumptions for a linear regression model are fulfilled?
We can form 95% confidence intervals around the estimated regression coefficients B1, B2, B3..
We can test the null hypotheses that B1= 0, B2= 0, B3= 0…
What is statistical modelling
The formalization of relationships between variables in the form of mathematical equations.
we infer the process by which data was generated!
theory-driven
What are statistical models, such as regression, typically used for?
Explanatory research to assess causal hypotheses that explain why and how empirical phenomena occur
Explanatory research usually infers from a random sample to an expected mean response of the underlying population.
We want unbiased estimates
In prediction modelling what are we not interested in?
inference and expectations
What is prediction modelling concerned with?
Reliable prediction of the outcome of unseen cases!
What does prediction modelling aim to minimise?
Minimizing the difference or “loss” between predicted and observed outcomes of new cases!
How can we estimate our model in prediction modelling?
By minimizing the error of new unseen cases and not the error of the original data set!
What would we like to know once we have developed our prediction model?
How well it predicts new unseen cases.
How can we can get a reliable estimate of prediction accuracy?
Predict outcome of new cases that were not used for model development
What is normal statistical modelling concerned with?
How well model predicts seen cases and thereby we calculate difference between observed and predicted and this error will be analyzed to see if it fulfills assumptions of normal distributed error
What does a prediction model usually perform better in?
Why is this a problem?
The sample used to develop the model (development or training sample) than in other samples, even if those samples are derived from the same population.
Problem as:
- We overestimate the predictive ability of our model
- Model performance is over-optimistic
1) What is something that works well for statistical modelling but not prediction modelling?
2) What is needed instead?
Assessing the assumptions of our model does not allow us to assume that it is going to work well on data that it has not seen before!
In other words, a high r2 (explained variance) in our training sample does not allow us to conclude that our model will have the desired performance if we apply it to new cases.
- Some kind of assurance of the accuracy of the predictions that our model is putting out! - need to validate our model.
To evaluate the performance of any prediction model, we need to test it on some unseen data.
Whether a model performs well or not is based upon what?
Models performance on unseen data