Psychometrics - Reliability + Coefficient Alpha Flashcards

1
Q

What is reliability?

A

The desired consistency or reproducibility of test scores

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

4 assumptions of classical test theory

A
  1. Each person has a true score that we could obtain if it weren’t for measurement error
  2. There is measurement error, but it’s random
  3. The true score of a person doesn’t change upon repeating tests, even though the observed score does
  4. The distribution of random errors will be the same for all ages
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Domain Sampling Model

A

The idea that we can’t construct a text that asks all possible questions within the domain being tested, so we have to select only certain ones. But using fewer items can lead to an introduction of error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Reliability analysis’s aim:

A

Establish how much error is made by using the score from the shorter test as an estimate of one’s true ability - error comes from multiple sources, and there are different ways of measuring reliability that are sensitive to different measurement errors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Four types of reliability

A
  1. Test-retest
  2. Parallel forms
  3. Internal consistency
  4. Inter-rater reliability
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What kind of error is the test-retest method designed for?

A

Time sampling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How does test-retest work?

A

You give someone the same test at two different points in time, and assess how much of difference there is in performance from the first test taking to the second

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Problems with test-retest reliability

A

Practice effect, testing effects, maturation, history

Also not ideal when you want to assess something that is expected to change over time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What source of error is parallel sampling designed to account for?

A

Item sampling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How does parallel forms reliability work?

A

You compose two different forms of the same test and get participants to do both

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Problems with parallel forms of reliability

A

How do we give both tests without having time problems? You need a bigger item pool. Testing effects.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What error does internal consistency reliability account for?

A

The reliability of one test administered on one occasion

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What does internal consistency measure and what three methods are used?

A

Do the different items within all measure the same thing to the same extent?

  1. Split half reliability
  2. Coefficient alpha
  3. KR-20
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

When do we use KR-20?

A

It’s used to find the alpha with dichotomous format measures

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How does split half reliability work?

A

A test is split in half, assessed and then correlated to see if the test is consistent.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Pros and cons of split half

A

+ You only need one test

  • It’s difficult to split the test
  • Halving the length of the test means it’s less reliable
  • Split half will change each time depending on which items sit in which test
17
Q

How do we account for the decrease in reliability that splitting the test will have?

A

Spearman-Brown correction!

18
Q

Formula for Spearman-Brown

A

Predicted reliability = 2 x correlation between halves / 1 + correlation between halves

19
Q

How do we account for the differences in Split Half depending on which items are where?

A

Cronbach’s Alpha!

20
Q

What does Cronbach’s alpha do?

A

It takes the average of all possible split half correlations for a test

21
Q

Formula for Cronbach’s alpha

A

A = (number of indicators)(average inter-item correlation)/(1 + (number of indicators - 1)(average inter-item correlation)

= kr/(1 + (k - 1)r)

22
Q

Alpha: correlation between number of items and reliability

A

Positive and non-linear, there is a rapid increase from 2-10, steady from 11-39 and it plateaus around 40.

23
Q

What levels of Cronbach’s alpha are we aiming for?

A
  1. 7 for exploratory research
  2. 8 for basic research
  3. 9 for applied scenarios
24
Q

What affects Cronbach’s alpha?

A
  • Multidimensionality
  • Bad test items
  • Number of items
25
Q

What source of error does inter-rater reliability help with?

A

Observer differences

26
Q

What does inter-rater reliability do?

A

It measures how consistently two or more rafters agree on rating something, based on the premise that multiple raters fro measurement can improve measurement reliability

27
Q

What inter-rater reliability do we use when?

A

2 raters = Cohen’s kappa

3+ = Fleiss’ kappa

28
Q

When do we use Cronbach’s alpha?

A

When we’re using raw scores

29
Q

When do we use standardized item alpha?

A

When scores have been standardized to account for things like age, gender etc.

30
Q

What is inter-item variance?

A

The variance between two items