Chapter 5 Flashcards

1
Q

Reliability (def)

A

consistency in measurement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Reliability coefficient

A

0 to 1 statistic.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

4 types of reliability coefficients

A

1) test-retest reliability

2) alternate-forms reliability

3) split-half reliability

4) inter-scorer reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Measurement error (textbook def)

A

Inherent uncertainty with any measurement, even after minimizing preventable mistakes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

2 influences that interfere with repeated measurement (in psych)

A

1) changes in object (eg. a constant flux of mood, alertness, motivation)

2) the act of measurement (i.e., carryover effects like fatigue, practice)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

“True Score”

A

not actually true to concept. True score is tied to the specific measurement instrument.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What ‘score’ measures the truth independent of measurement?

A

Construct score.

the underlying score of some construct (eg. depression)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

variance is made of what two subtypes of variance?

A

True variance (actual differences between people?) + Error variance (random variances that are irrelevant)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Define reliability in terms of variance

A

Proportion of total variance attributed to true variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Random vs. Systematic Error

A

Random: unpredictable, inconsistent, without pattern

Systematic: predictable, constant, can be adjusted for

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Bias (in error)

A

The degree of systematic error that influences measurement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How does item/content sampling contribute to error variance?

A

The specific content in some test may affect the results (eg. i hope they ask this question and not this)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What test administration effects contribute to error variance?

A

Environment: war, heat, gum, pencil, etc.

Testtaker variables: lack of sleep, emotions, drugs, etc.

Examiner-related variables: physical appearance, presence/absence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How does test scoring and interpretation contribute to error variance?

A

Some subjectivity in certain tests (eg. essays, creativity, etc.) can influence measurement.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

test-retest reliability coefficient is also called what?

A

Coefficient of stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What might affect test-retest reliability estimates?

A

Experience, practice, memory, fatigue, etc. may intervene.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

alternate-forms/parallel-forms reliability estimates coefficient name

A

Coefficient of equivalence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Parallel vs. Alternate forms reliability

A

Parallel forms: Means and variances of test scores are equal

Alternate forms: different versions of same test, but aren’t parallel

19
Q

2 similarities between parallel/alternate and test-retest reliability

A

1) two test administrations with same group

2) test scores can be affected by factors like fatigue, practice, learning, etc.

20
Q

What additional source of error variance is present in alternate/parallel-forms reliability?

A

Item/Content sampling

21
Q

Split-half reliability

A

Correlating two pairs of scores from a single test.

one half of a test Pearson r with another half, then adjust with Spearman-Brown formula

22
Q

Odd-even reliability

A

split-half reliability by using odd vs. even numbers

23
Q

How do number of items affect reliability coefficient? What method can see how many items needed?

A

Spearman-Brown.

More items is more reliability

24
Q

What coefficient for inter-item consistency?

A

Coefficient alpha

25
Q

Inter-scorer reliability

What coefficient?

A

Degree of consistency between 2 or more scorers.

Coefficient of inter-scorer reliability

26
Q

DSM-5 Inter-rater reliability

A

Kappa = 0.44 (fair level moderately greater than chance)

27
Q

Transient error

A

Error due to testtaker’s feelings, moods, or mental state over time

28
Q

Homogeneity vs. Heterogeneity of test items

A

Homogenous: Functionally uniform items. Measures one factor (eg. one ability/trait). High internal consistency should happen

Heterogenous: Not just one factor measured in the test.

29
Q

Does high internal consistency mean homogeneity of items?

A

Not necessarily.

More items will lead to high internal consistency coefficients as long as they’re positively correlated

30
Q

Dynamic vs. static characteristics

A

Dynamic: Presumed to be relatively situational and changing

Static: presumed to be relatively unchanging

31
Q

Restriction/Inflation of range

A

When some subgroup inflates or restricts the correlational analysis??

32
Q

Power Test

A

Enough time to attempt all items, but so difficult that nobody gets perfect score

33
Q

Speed test

A

Same level of difficulty in items and testtakers should complete everything correctly if unlimited time.

But only some will be able to complete the whole test

34
Q

What’s differences in assumptions between CTT and IRT? (not specific, but ya..)

A

CTT assumptions are weak and easily met. IRT are rigorous.

35
Q

Domain Sampling Theory

A

Reliability is based on how well a score assesses the domain of where a sample is drawn.

36
Q

What is universe score in generalizability theory?

A

The true score (given same conditions, the same score will be obtained)

37
Q

Generalizability Study

Coefficient of generalizability

A

how generalizable scores from a particular test are if administered in different situations.

38
Q

Decision study

A

Usefulness of test scores in helping user make decisions. Follows generalizability study

39
Q

Another way to say Item response theory

A

Latent-trait theory

40
Q

Within CTT, what is the weight assigned to each item on a test?

A

Equal weight. IRT is differentital weight.

41
Q

Dichotomous test items

A

can only answer with one of two responses

42
Q

Polytomous test items

A

3 or more alternative responses

43
Q

Rasch Model

A

a type of IRT model with underlying distribution assumption

44
Q

Which measure is used to compare differences between scores?

A

Standard error of the difference