RELIABILITY Flashcards

1
Q

involves the consistency of the measuring tool: the precision with which the test measures and the extent to which error is present in measurements.

A

RELIABILITY

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

is an index of reliability, a proportion that indicates the ratio between the true score variance on a test and the total variance.

A

reliability coefficient

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Sources of Error Variance

A
  • Test construction
  • Test administration
  • Test scoring and interpretation
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Sources of Error Variance

item sampling or content sampling

A

Test construction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Sources of Error Variance

Test administration:

A
  • test environment
  • test-taker variables
  • examiner-related variables
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Test Administration:

The room temperature, the level of lighting, and the amount of ventilation and noise,

A

Test Environment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Test Administration:

Pressing emotional problems, physical discomfort, lack of sleep, and the effects of drugs or medication.

A

Test-taker variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Test Administration:

Physical appearance and demeanor

A

Examiner-related variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Sources of Error Variance

Scorers and scoring systems

A

Test scoring and interpretation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Reliability Estimates

A
  • Test-retest reliability
  • Parallel-Forms and Alternate-Forms Reliability Estimates (coefficient of equivalence)
  • Split-Half Reliability Estimates
  • KR-20 AND COEFFICIENT ALPHA FORMULA
  • Measures of Inter-Scorer Reliability
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Reliability Estimates

Estimate of reliability obtained by correlating pairs of scores from the same people on two different administrations of the same test.

A

Test-retest reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

tendencies to act, think, or feel in a certain manner in any given circumstance

A

TRAIT

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Test-retest reliability

When the interval between testing is greater than six months, the estimate of test-retest reliability is often referred to as the __

A

coefficient of stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Parallel-Forms and Alternate-Forms Reliability Estimates

What type of reliability estimate is obtained when two different versions of a test are constructed to be parallel?

A

Alternate-Forms Reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Parallel-Forms and Alternate-Forms Reliability Estimates

What type of test reliability exists when the means and variances of observed test scores are equal for each test form?

A

Parallel-Forms Reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Parallel-Forms and Alternate-Forms Reliability Estimates

What are the two ways to obtain an estimate of parallel-forms reliability?
Answer:

A
  1. Administering two test forms to the same group
  2. Considering factors like motivation, fatigue, and practice effects
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Parallel-Forms and Alternate-Forms Reliability Estimates

What is the primary source of error variance in alternate- or parallel-forms reliability?

A

Item Sampling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Parallel-Forms and Alternate-Forms Reliability Estimates

What is one drawback of parallel-forms reliability testing due to its complexity and cost?

A

It is time-consuming and expensive

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Parallel-Forms and Alternate-Forms Reliability Estimates

What type of reliability can be obtained without developing an alternate test form or administering a test twice?

A

Internal Consistency Reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Reliability Estimates

What reliability estimate is obtained by correlating two pairs of scores from equivalent halves of a single test?

A

Split-Half Reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Split-Half Reliability

What is the first step in obtaining a split-half reliability estimate?

A

Divide the test into equivalent halves

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Split-Half Reliability

What statistical method is used to calculate the correlation between two halves of a test?

A

Pearson r

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Split-Half Reliability

Which formula is used to adjust the half-test reliability in split-half reliability estimates?

A

Spearman-Brown Formula

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Split-Half Reliability

Why is dividing a test in the middle not recommended for split-half reliability testing?

A

It may not create equivalent halves

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

Split-Half Reliability

What are three acceptable ways to split a test for split-half reliability estimation?

A
  1. Randomly assign items
  2. Assign odd-numbered items to one half and even-numbered items to the other
  3. Divide the test by content and difficulty
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

Split-Half Reliability

Which formula estimates the reliability of a test when its length is changed?

A

Spearman-Brown Formula

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

KR-20 and Coefficient Alpha Formula

What does inter-item consistency measure in a test?

A

The degree of correlation among all test items

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

KR-20 and Coefficient Alpha Formula

What term refers to a test that measures a single trait, leading to higher inter-item consistency?

A

Homogeneity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

KR-20 and Coefficient Alpha Formula

What term describes a test that measures different factors, leading to lower inter-item consistency?

A

Heterogeneity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

KR-20 and Coefficient Alpha Formula

Which formula is used to determine the inter-item consistency of dichotomous items?

A

Kuder-Richardson Formula 20 (KR-20)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

KR-20 and Coefficient Alpha Formula

Which formula is used to assess the internal consistency of tests with non-dichotomous items, such as Likert scales?

A

Coefficient Alpha

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

KR-20 and Coefficient Alpha Formula

Which formula provides the mean of all possible split-half correlations?

A

Coefficient Alpha

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

Reliability Estimates

What is another term for scorer reliability, which measures the consistency between different judges or raters?

A

Inter-Rater Reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

Measures of Inter-Scorer Reliability

What does inter-scorer reliability assess in a test?

A

The degree of agreement between two or more scorers, judges, or raters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

What statistical value is used to measure inter-scorer reliability?

A

Coefficient of Inter-Scorer Reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

Nature of Tests

What type of test items measure a single ability or trait and have a high degree of internal consistency?

A

Homogeneous Items

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

Nature of Tests

What type of test items measure multiple abilities or traits, leading to lower internal consistency estimates?

A

Heterogeneous Items

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

Nature of Tests

Which type of test items typically result in a high internal consistency reliability estimate?

A

Homogeneous Items

39
Q

Nature of Tests

For which type of test items is test-retest reliability a more appropriate measure than internal consistency?

A

Heterogeneous Items

40
Q

Nature of Tests

What type of characteristic is presumed to be ever-changing due to situational and cognitive experiences?

A

Dynamic Characteristic

41
Q

Nature of Tests

What type of characteristic is presumed to be relatively unchanging over time?

A

Static Characteristic

42
Q

Nature of Tests

A person’s mood, which changes depending on experiences and situations, is an example of what kind of characteristic?

A

Dynamic Characteristic

43
Q

Nature of Tests

A person’s fingerprint, which remains consistent throughout life, is an example of what kind of characteristic?

A

Static Characteristic

44
Q

Nature of Tests

What occurs when a test limits the variability of scores, potentially underestimating the true relationship between variables?

A

Restriction of Range

45
Q

Nature of Tests

What happens when a test expands the variability of scores, potentially exaggerating the true relationship between variables?

A

Inflation of Range

46
Q

Nature of Tests

A
  1. Homogeneity vs Heterogeneity of test items
  2. Dynamic vs Static characteristics
  3. Restriction or inflation of range
  4. Speed tests versus power tests
  5. Criterion-referenced tests
47
Q

it provides an estimate of the amount of error inherent in an observed score or measurement

A

Standard Error of Measurement (SME)

Also known as the standard error of a score

48
Q

is a judgment or estimate of how well a test measures what it purports to measure in a particular context

49
Q

VALIDITY

Three Categories of Validity:

A
  1. Content Validity
  2. Criterion-Related Validity
  3. Construct Validity
50
Q

VALIDITY

Which type of validity is concerned with evaluating the actual content of a test to ensure it represents what it is supposed to measure?

A

Content Validity

51
Q

VALIDITY

Which validity category involves comparing test scores with other established measures to determine its accuracy?

A

Criterion-Related Validity

52
Q

VALIDITY

Which type of validity focuses on analyzing how test scores relate to a theoretical framework or concept?

A

Construct Validity

53
Q

VALIDITY

If a math test is checked to ensure it includes all necessary topics, which type of validity is being assessed?

A

Content Validity

54
Q

VALIDITY

If a new depression scale is compared to an existing clinical measure of depression, which type of validity is being examined?

A

Criterion-Related Validity

55
Q

VALIDITY

A psychologist analyzes whether a personality test aligns with established personality theories. What type of validity is this?

A

Construct Validity

56
Q

VALIDITY

What type of validity refers to how much a test appears to measure what it is supposed to, from the perspective of the test taker?

A

Face Validity

57
Q

VALIDITY

What is the term for a judgment about how relevant and appropriate test items seem to be?

A

Face Validity

58
Q

VALIDITY

Which type of validity can influence a test taker’s motivation and cooperation based on their perception of the test’s effectiveness?

A

Face Validity

59
Q

VALIDITY

What type of validity assesses how well a test represents the entire domain of behavior it is designed to measure?

A

Content Validity

60
Q

VALIDITY

What is a method developed by C. H. Lawshe to measure agreement among judges on the importance of test items?

A

Content Validity Ratio (CVR)

61
Q

VALIDITY

What type of validity assesses how well test scores can predict an individual’s performance on a related measure of interest?

A

Criterion-Related Validity

62
Q

VALIDITY

What is the standard against which test scores are evaluated in criterion-related validity?

63
Q

VALIDITY

What error occurs when a rater’s knowledge of test scores influences their ratings?

A

Criterion Contamination

64
Q

VALIDITY

What are the characteristics of a good criterion?

A

a. relevant significance
b. valid measure
c. uncontaminated obsolete

65
Q

VALIDITY

Two Types of Validity

A

a. Concurrent Validity
b. Predictive Validity

66
Q

VALIDITY

Which type of validity measures the relationship between test scores and a criterion measure obtained at the same time?

A

Concurrent Validity

67
Q

VALIDITY

Which type of validity assesses how well a test predicts future performance on a criterion measure?

A

Predictive Validity

68
Q

VALIDITY

What is the correlation coefficient that indicates the relationship between test scores and the criterion measure?

A

Validity Coefficient

69
Q

VALIDITY

What type of validity evaluates whether test scores accurately represent an abstract concept or theoretical idea?

A

Construct Validity

70
Q

VALIDITY

What is an informed, scientific concept developed to describe or explain a behavior, such as motivation or depression?

71
Q

VALIDITY

Which type of validity requires formulating hypotheses about how high and low scorers should behave?

A

Construct Validity

72
Q

Construct Validity

Evidence of Construct Validity

A
  • Evidence of homogeneity
  • Evidence of changes with age
  • Evidence of pretest–posttest changes
  • Evidence from distinct groups
  • Convergent evidence
  • Discriminant evidence
73
Q

Evidence of Construct Validity

What type of evidence supports construct validity by showing that a test measures a single concept?

A

Evidence of Homogeneity

74
Q

Evidence of Construct Validity

What type of evidence supports construct validity when test scores change with age as expected?

A

Evidence of Changes with Age

75
Q

Evidence of Construct Validity

Which type of evidence is based on changes in test scores after an intervention or over time?

A

Evidence of Pretest–Posttest Changes

76
Q

Evidence of Construct Validity

What type of construct validity evidence shows that test scores vary as expected among distinct groups?

A

Evidence from Distinct Groups

77
Q

Evidence of Construct Validity

Which type of evidence is shown when test scores correlate well with other measures of the same construct?

A

Convergent Evidence

78
Q

Evidence of Construct Validity

What type of evidence demonstrates that a test does not correlate with measures of unrelated constructs?

A

Discriminant Evidence

79
Q

TEST BIAS

Different Kinds of Test Bias

A
  • Rating Error
  • Halo Effect
  • Horn Effect
  • Contrast Error
  • Recency Bias
80
Q

TEST BIAS

3 TYPES OF RATING ERROR

A

a. Leniency error or generosity error
b. Severity error
c. Central tendency error

81
Q

RATING ERROR

What type of rating error occurs when a rater is overly forgiving in scoring, marking, or grading?

A

Leniency Error

(Generosity Error)

82
Q

RATING ERROR

A teacher gives almost all students high grades, even if some did poorly on their tests. What kind of rating error is this?

A

Leniency Error

(Generosity Error)

83
Q

RATING ERROR

What type of rating error occurs when a rater is overly harsh in scoring?

A

Severity Error

84
Q

RATING ERROR

A supervisor consistently gives employees low performance scores, even if they meet expectations. What kind of rating error is this?

A

Severity Error

85
Q

RATING ERROR

What type of rating error happens when a rater avoids extreme ratings and tends to score in the middle of the rating scale?

A

Central Tendency Error

86
Q

RATING ERROR

A teacher grades all students between 80-85, even though some deserve a 95 and others a 70. What kind of rating error is this?

A

Central Tendency Error

87
Q

TEST BIAS

What is the tendency to give a higher rating than deserved due to the failure to distinguish between different aspects of a person’s behavior?

A

Halo Effect

88
Q

TEST BIAS

A manager gives an employee excellent ratings in all areas because they have a friendly personality, even though their actual work performance is average. What kind of bias is this?

A

Halo Effect

89
Q

TEST BIAS

What bias occurs when one negative aspect of performance influences all other ratings, resulting in an overall lower score?

A

Horn Effect

90
Q

TEST BIAS

A teacher gives a student low marks in all subjects because the student misbehaved in class, even though they perform well academically. What kind of bias is this?

A

Horn Effect

91
Q

TEST BIAS

What type of rating error occurs when raters compare individuals to each other instead of evaluating them against performance standards?

A

Contrast Error

92
Q

TEST BIAS

A recruiter rates an average candidate poorly because they interviewed right after an exceptional candidate. What type of error is this?

A

Contrast Error

93
Q

TEST BIAS

What bias occurs when a leader bases their evaluation on an employee’s most recent performance rather than their overall performance?

A

Recency Bias

94
Q

TEST BIAS

A manager rates an employee poorly because they made a mistake last week, even though they performed well throughout the year. What kind of bias is this?

A

Recency Bias