Week 9-Reliability and Validity Flashcards
What Makes a Good Experiment?
■ Reliability.
■ Sensitivity.
■ Validity.
What’s Reliability?
■ Consistency/dependability of a measure.
■ The ability of a measure to produce the same or similar results on repeated administrations.
■ Is it any good at measuring?
■ Internal reliability
■ External Reliability
What’s Internal Reliability?
■ The extent to which a measure is consistent
within itself.
■ Split-half reliability:
– Compares the results of one half of a test with the other half.
– Split the items into two groups and correlate
them.
– A participant scoring high on one half, should
score high on the other half.
But what if all the good questions were at the beginning, rubbish questions at the end?
■ May give a weaker correlation than is fair.
■ Any correlation might just be down to chance, depending on how the questionnaire was split.
What’s Cronbach’s alpha?
■ A test that splits items equally in every way
possible.
■ It then correlates ALL halves with ALL other halves
■ Much more robust measure of internal reliability.
■ Should have a high correlation coefficient >.70
■ SPSS also gives information about how each item
correlates with all other items.
■ And how alpha would change if an item was
removed.
■ The higher the alpha – the more reliable the questionnaire.
What’s External Reliability?
■ The extent to which a measure varies from one use to another.
■ Test-retest reliability:
– the stability of a test over time.
– A good test is consistently reliable.
– Administer the test now, then give the same test later to the same participants
– A good test will have a high correlation.
What’s inter-rater reliability (external reliability)
– Usually used in observational studies
– Degree to which different raters give consistent estimates of the same behaviour.
– Correlation to check reliability.
■ To improve inter-rater reliability:
– Clear categories/definitions.
– Training.
How can you improve reliability in general?
– Improve quality of items.
– Increase/decrease number of items.
– Increase sample size.
– Choose appropriate sample.
– Control conditions.
What’s sensitivity?
■ Detecting even a small effect of the IV on the DV.
– Large sample size
– Sizeable/varied effects
– Control variability
What’s the benefits of sensitivity?
■ Not too hard or too easy.
■ Wide range of scores.
■ E.g. The effect of high levels of alcohol on reading ability?
– IV = alcohol intake.
– DV?
– Ask participants if they found the words hard to
read - Yes/No.
– Time them reading the word list.
■ Properties of your sample:
– Facebook/instagram usage – students or
pensioners?
What’s Validity?
■ A test/measure.
■ A study/experiment.
■ Test validity - the ‘truthfulness’ of a measure in that it measures what it claims to.
– Content validity.
– Construct validity.
– Criterion validity.
What’s content validity?
– Face validity - whether the test appears to measure what it claims to.
– Purpose is obvious.
– Social desirability?
What’s construct validity?
The degree to which a test measures the construct/psychological concept at which it is aimed.
– Convergent validity
– The degree to which it correlates with other
measurements assessing the same construct.
– Divergent/Discriminant validity
– The degree to which it does not correlate with
other measurements assessing different concepts.
What’s the criterion validity?
Whether a test reflects a certain set of abilities i.e. the degree to which a measurement can accurately predict specific criterion variables.
– Concurrent validity:
– How well a test correlates with a previously
validated measure, given at the same time.
– Predictive validity:
– How well it predicts future performance.
The validity of a study/experiment:Types
– External validity
– Internal validity
– Ecological validity