Psychometrics: reliability Flashcards
What is a reliable test?
- consistency in measurement
- the precision with which the test score measures achievement
What is reliability
- the desired consistency or reproductibility of test scores (does it give me the same accurate measurement each time it is used?)
- no test is free from error
Reliability formula
x=T+e
x- the observed score
T- the true score
e- the error
The Four Assumptions of Classical Test Score Theory
- Each person has a true score we could obtain if there was no measurement error
- there is measurement error- but this error is random
- the true score of an individual doesnt change with repeated applications of the same test, even though their observed score does
- the distribution of random errors and thus observed test scores with be the same for all people
Standard Error of measurement (SEM)
-works out how much measurement error we have by working out how much on average, an observed score on our test differs from the true score
(standard deviation)
Problems with Classical Test Score Theory
- Population dependent
- Test dependent
- Assumption of equal error measurement
Domain Sampling Model
- a central concept of Classical Test Theory
- cant ask all possible questions on a test so only use a few test items (sample)
- using fewer test items can lead to the intro of error
- as sample gets larger, estimate is more accurate
4 Types of reliability
- Test-retest reliability
- Parallel forms reliability
- Internal consistency
- inter-rater reliability
Test-retest reliability
- give someone a test and then give them the same test later on
- if scores are highly correlated, we have a good test-retest reliability
- correlation between 2 scores = co-efficient of stability
- time sampling
Issues with test-retest
- . can it be used when measuring mood/stress?
- scores increase because done them before
- if thing being measure changes?
- what if an event happens between tests administrations to change the thing being tested?
Parallel forms reliability
- 2 forms of the same test (questionnaire with different items)
- correlation between the two = co-efficient of equivalence
- item sampling
Ways to change test in parallel forms reliability
- question response alternatives are reworded
- order is changed
- change wording of question
Issues with parallel forms reliability
- what if different forms are given at two different times?
- do you give the form to the same or different people?
- what if people work out how to answer the one form from doing the other form?
- do you have two forms of the test and/or do we want to develop two forms of the same test?
Internal Consistency
-do different items within a test all measure the same thing, to an extent?
Examples of internal consistency tests
- split-half reliability
- KR20
- coefficient alpha