Validation Flashcards
What is validation?
Checking to see how good a model is
what is a real effect?
real relationship between attribute and response. same in all datasets
what is a random effect?
random, but looks like real effect. different in all datasets
Which effects do you fit your model on?
Both real and random effects. New data will have fewer correct predictions because it only has the real effects in common with the our fit, not random effects
How do you measure a models performance?
Split the data - larger set of data to fit the model and smaller set to measure the models effectiveness
How do you measure multiple models performance? And why?
train/test/validation split
-test the models on the validation split. evaluate the individual model with the test split
-high performing models are more likely to have above average random effects
-if all models were equally good, the only performance we would measure is random effects
When do you use a validation and test set and when do you use only test?
if you’re chooosing between models choose the best model using validation set. then estimate quality with test set. otherwise estimate quality with test set
How much data goes into each split?
1 model
-70 - 90% for training
comparing models
-50-70% for training and split the rest between test and validation
what is Random splitting data and what is the downside
randomly choose the datapoints for each group. could give one set more of a certain type of data than the other
rotation splitting of data
take turns selecting points. 1st goes to training, 2nd to test, and 3rd to validation. then repeat. could introduce bias if your data is structured
What problem does cross validation solve?
important points only show up in the training or test set
what is cross validation?
split the data evenly. then test each part against a model trained on all others
what is k-fold cross validation
for each of the k parts:
train the model on all other parts
evaluate it on the one remaining part
avg. the k evaluations to estimate the models quality.
no standard number but k = 10 is common
With crossfold validation, which model do you pick as your final model?
None. You train a final model on all the data whose accuracy is the avg of the k tests