Reliability Chapter 4 Flashcards

1
Q

The Domain sampling model definition of reliability

A

Reliability is the relation of the variance of the observed score on a shorter test in the variance of the long run true score

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Domain sampling model how is error introduced:

Problem and solution

A

By using a sample of items rather than the entire Domain of knowledge. The greater the items the higher the reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Classical test score Theory

Formula

A

Observed score equals true score plus error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Classical test Theory assumption

A

Error is random true score of a person doesn’t change with repeated applications of a test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Item response Theory explanation

A

Computer chooses a the questions to give an individual based on how many correct questions they have already got question difficulty goes up and down based on the individuals getting items of similar difficulty right or wrong

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Item response Theory reliability

A

A shorter test has higher reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Sources of error

2 things.

A

Situational factors EG loud noises and temperature

Items not representative of the Domain

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Three ways to estimate test reliability

A

Test-retest method
Parallel forms
Internal consistency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Time sampling the test retest method

A

Administer the test on two occasions correlate the scores

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Problems with test-retest

2 things.

A

Carry over effects
practice effects
Maturation effects
Length of interval between test is important

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Parallel forms methods

A

2 equivalent forms that measure the same atribute using different items
Sometimes given on the same day with counterbalanced order
One of the most rigorous assessments of reliability commonly in use

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Parallel forms methods

A

2 equivalent forms that measure the same atribute using different items
Sometimes given on the same day with counterbalanced order
One of the most rigorous assessments of reliability commonly in use

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Parallel forms methods

A

2 equivalent forms that measure the same atribute using different items
Sometimes given on the same day with counterbalanced order
One of the most rigorous assessments of reliability commonly in use

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Split half method of reliability:
Definition
Problem
Solution

A

Two halves of a test
Correlatio

The spearman-brown formula corrects for using a shorter test
Estimates what the correlation between the two halves would have been if each half had been the length of the whole test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Internal consistency reliability split half how to overcome the problem that all halves are slightly different in difficulty

A

Kuder Richardson formula considers always of splitting a test in half

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Split half reliability when to use cronbachs Alpha:

2 reasons

A

When variance is between the two halves is unequal

Scores have more options than right or wrong

16
Q

Reliability of a difference score:

Definition and how to find it?

A

If you want to know the difference between performances at 2 points in time
Convert scores to Z units
Reliability for the difference score is lower than the reliability of the first and second measures

17
Q

Reliability and behaviour observation studies:
Names for it
Best formula and why

A

Inter-rater reliability also called interscorer interobserver or into judge
Best method Kappa statistic
Kappa for chance agreement the servers are rating whether a behaviour occurred or did not occur there is a 50% chance of agreement

18
Q

Standard errors of measurement and the rubber yardstick: problem and solution

A

The larger the standard error of measurement the less certainty about the accuracy with which the atribute was measured.
Researchers use confidence intervals to show the probability that the true score Falls within a certain interval around the observed score

19
Q

How reLiable is reliable for basic research and for making decisions about people?

A

Basic research.70-.80

Instruments used to make important decisions .95

20
Q

Increasing reliability of a test

Definition and formula

A

Increase the number of items. The prophecy formula can estimate how many more items a test needs to gain an acceptable level of reliability
Prophecty Formula assumptions the probability of error for new items = the probability of error for original items tiredness and boredom is not going to increase error

21
Q

Discriminabilty analysis

A

Correlation between performance on single item and total test score if low item probably measures something different from the others.
Item may be too easy

21
Q

Attenuation of correlation

A

Correlations are diminished by measurement error

22
Q

Relibility needs Factor and item analysis because….

A

Tests are reliable if they are uni-dimensional one Factor should account for more variance than any other Factor

23
Q

Correcting for attenuation

A

To Estimate what the correlation would have been between two measures without error
Find the reliability of both measures and the correlation between them
Use the correction for attenuation reliability formula

24
Q

Use Krs20

A

Dichotomous

25
Q

Use alpha

A

Likert skewed batiance

No right or wrong answer

26
Q

Use pearson

A

Two continuous variables