Chapter 4: Reliability Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

What is reliability?

A

How often you get the same result, think of a dart board, MEASURING CONSISTENCY

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What contributes to measurement error?

A

Situational
- Test environment
- Test-taker
- Examiner

Test Creation
- Item sampling
- Test scoring: criteria, administrator bias
- Test interpretation: arbitrary, low inter-rater agreement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What components make up Classical Test Score Theory?

A

assumes that each person has a true score, T, that would be obtained if there were no errors in measurement.

x (observed score) = T (True Score) + E (Error)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

In what ways can error impact the observed score?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Test reliability is usually estimated in one of what three ways? Know the major concepts in each way.

A
  • test-retest: do the scores stay consistent across when administered again?
  • internal consistency: how well do items test across test-takers with a lot of ability and those without ability
  • parallel forms: two tests that measure the same-thing
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is a carryover effect?

A

When the test participant can remember items from their previous time taking the test, and thus, may score differently.
When the first session influences the second

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Define parallel/alternate forms reliability. What are its advantages and disadvantages?

A

Parallel forms: When you have two tests that are different from each other that measure the same thing.

Disadvantage: Extremely hard to do/construct

Advantage: Reduce memory bias, you can re-administer a test, decrease cheating

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Define split half reliability. How is this measured?

A
  • Give once and divide into separately scored halves
  • Correlate the halves
  • Only requires one administration
  • Deflated reliability
  • Difficult to split into equivalent halves
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How do the different aspects of internal consistency differ?

A

Measures whether all the items are measuring the same thing

Split-alf:
Spearman-Brown:
KR20:
Cronback’s alpha

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Understand the major components of inter-rater reliability.

A
  • Measured using Kappa coefficient
  • Different raters will rate the items differently
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the Kappa statistic and how does it relate to reliability?

A
  • measures inter-rater reliability
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Summary of Reliability

A

Error and Reliability
Difference between observed and true score

Classic test score theory
x= t + e

Evaluating sources of error
Test-retest, alternative/parallel form, split-half, internal consistency, inter-rater

Fixing reliability
Increase items, drop unhelpful items, correct for measurement error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Know the Summary of Reliability Table from lecture

A

https://docs.google.com/presentation/d/1tvNLDV2q4N0Rn0JajLM5sBhzXSxXrbh7/edit#slide=id.p29

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What does the standard error of measurement do?

A

If it changes randomly, its distribution should be normal
The mean will be an estimate of the true score
The standard deviation will be the standard error of measurement (how much the score varies from the true score)
Assumed that the standard errors are the same for everyone
True variance is ideal, the ACTUAL differences in ability before error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What factors should be considered when choosing a reliability coefficient?

A

Depends of the nature of the test.

For example, if you have raters rating an essay then calculate interrater reliability. If you are looking to see test items consistently rate those with low and high ability use internal consistency measures.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Why types of irregularities might make reliability coefficients biased or invalid?

A

Inflated variance, constricted variance, outliers

17
Q

How can one address/improve low reliability?

A

Increase items, drop unhelpful items, correct for measurement error

18
Q

What is the purpose of factor and item analysis?

A

a powerful data reduction technique that enables researchers to investigate concepts that cannot easily be measured directly

19
Q

What example was given in class regarding reliability

A

dart board