Psychological Testing & Assessment Flashcards

1
Q

What are the 7 assumptions about psychological testing and assessment?

A

Psychological Trait & States Exist.
These can be quantified and measured.
Test behavior predicts non-test behavior.
Tests have strengths and weaknesses
Sources of error are part of the process.
Testing & assessment can be done in a fair and unbiased way.
Testing & assessment benefit society.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Any distinguishable, relatively enduring way in which individual varies from one another.

A

Trait

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Distinguishes one person from another but relatively less enduring.

A

State

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

The key to insight.

A

Item content

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Long standing assumption that factors other than what a test attempts to measure will influence performance on the test.

A

Error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Refers to the component of test score attributable to sources other than the trait or ability measured.

A

Error variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

The stability or consistency of measurement.

A

Reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Elements of error

A

Observer’s error
Environmental changes
Participant’s changes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Reliability coefficient should not go beyond _.

A

+/- 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

The __ the coefficient alpha, the higher the reliability.

A

Higher

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Test scores gain reliability as the number of _ increases.

A

Items

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

It means dependable, consistency or stability.

A

Reliability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Defined as one on which test takers fall in the same positions relative to each other.

A

Reliable test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Reliability assumes that test scores reflect 2 factors which are:

A

True characteristics
Random measurement of error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Stable characteristics of the individual.

A

True characteristics

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Chance features of the individual or the situation.

A

Random measurement of error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Tools used to estimate or infer the extent to which an observed score deviates from a true score.

A

Standard error of estimates/measurement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Mathematical representation of Random Measurement error

A

X= T+E

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

In a reliable test, the value of E should be close to _ and the value of T should be close to the _.

A

0
Actual test score X.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Formula for the proportion of test score reflecting a person’s true characteristics.

A

T/X

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Formula for the proportion of test score reflecting random error.

A

E/X

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

The reliability of the test is actually based on _.

A

Performance of people

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Difference between people in test scores reflect differences between them in _ plus differences in the effect of _ factors.

A

Actual knowledge/characteristics
Chance factors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What are the sources of error variance?

A

Test construction
Test administration
Test scoring and interpretation
Other sources of error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

It correlates performance on two interval scale measures and uses that correlation to indicate the extent of true score differences.

A

Reliability Analysis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

Used to evaluate the error associated with administering a test at two different times.

A

Test-retest method

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

The test retest method is of value only when we measure traits or characteristics that _.

A

Do not change overtime.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

Ideally, test-retest method should have _ months or more interval.

A

6

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

When the interval between testing is greater than 6 months.

A

Coefficient of stability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

It compares two equivalent forms of a test that measures the same attribute.

A

Parallel/Alternative forms method

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

The 2 forms in parallel/ alternative forms method use _ items while the rules used to select items of a particular difficulty level are _. (Same/Different)

A

Different
Same

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

A test is given to some and divided into halves that are scored separately.

A

Split half method

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

One subscore is obtained for the odd number items in the test and another for the even-numbered items. Used in split half method.

A

Odd-even system

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

Allows you to estimate what the correlation between the 2 halves would have been if each half had been the length of the whole test.

A

Spearman-Brown Formula

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

Refers to the degree of correlation among all the items on a scale.

A

Inter-item consistency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

Inter item consistency is calculated using _ administration of a _ test form. (frequency)

A

Single

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

Inter item consistency is useful in assessing the _ of a test.

A

Homogeneity

38
Q

What are the methods used to obtain estimates of interval consistency?

A

KR 20
Cronbach alpha

39
Q

Cronbach developed a more general reliability estimate which he called _.

A

Cronbach alpha

40
Q

Cronbach alpha typically range from _ to _, _ values are theoretically impossible.

A

0-1
Negative values

41
Q

The degree of agreement or consistency between two or more scorers with regard to a particular measure.

A

Inter-scorer reliability

42
Q

Inter-scorer reliability is for what kinds of test?

A

Creativity or projective test

43
Q

Judgment or estimate of how well a test measures what it purports to measure in a particular measure.

A

Validity

44
Q

A judgment based on evidence about the appropriateness of inferences drawn from rest scores.

A

Validity

45
Q

Process of gathering and evaluating evidence about validity.

A

Validation

46
Q

Validation process if test users plan to alter format, instruction, language or content of the test. Example is the national standardized test to braille

A

Local validation.

47
Q

As reliability of the test increases, the highest possible value of _ increases.

A

Validity coefficient

48
Q

What are the 5 measures of reliability?

A

Test retest method
Parallel/Alternative forms method
Split-half method
Inter-item consistency
Inter-scorer Reliability

49
Q

The first theory behind validity analysis is that the true score component reflects factors producing _ in test scores while E, error reflects factors producing _ in test scores.

A

Stability
Instability

50
Q

The second theory behind validity analysis is focused specifically on the variable producing _ differences.

A

True score differences.

51
Q

What are the 2 components of true score?

A

Stable characteristics of individual relevent to the purpose of the test.
Stable characteristics of individual irrelevant to the purpose of the test.

52
Q

Formula of systematic measurement error.

A

R+I = T

53
Q

Formula of both error components (systematic and random error)

A

X= R + I + E

54
Q

When the item looks like they measure what they are supposed to measure.

A

Face Validity

55
Q

The judgment about the item appropriateness in face validity is made by _.

A

Test takers

56
Q

It is important whenever a test is used to make inferences about the broader domain of knowledge or skills represented by a sample of items.

A

Content validity

57
Q

Content validity is important to what kinds of performance test?

A

Maximal and Typical Performance Test

58
Q

The ability of a test to predict performance in another measure.

A

Criterion Validity

59
Q

In criterion Validity, the test is referred to as the _ labeled X and the validation measure as the _ labeled Y.

A

Predictor;
Criterion

60
Q

It is important whenever a test is used to make decisions by predicting future performance.

A

Criterion Validity

61
Q

A judgment of how adequately a test score can be used to infer an individual’s most probable standing on some measure of interest- being the criterion.

A

Criterion- related Validity

62
Q

What are the 2 types of validity evidence?

A

Predictive validity
Concurrent validity

63
Q

An index of the degree to which a test score predicts some criterion measure. Type of validity evidence.

A

Predictive validity

64
Q

An index of the degree to which a test score is related to some criterion measure obtained at the same time (concurrently)

A

Concurrent validity

65
Q

Used whether a test measures what it is intended to measure. Referred to as the personality dimension of personality traits.

A

Construct validity

66
Q

What are the 2 construct validation techniques?

A

Congruent validity
Discriminant or divergent validity

67
Q

Construct validation technique that measures correlating the same construct.

A

Congruent validity

68
Q

Construct validation technique that measures correlating it to inconsistent constructs.

A

Discriminant or divergent validity

69
Q

An informed, scientific idea developed or hypothesized to describe or explain behavior.

A

Construct

70
Q

What are the 5 types of validity?

A

Face validity
Content validity
Criterion Validity
Criterion- related Validity
Construct validity

71
Q

What are the statistical test bias?

A

Intercept bias
Slope bias

72
Q

A judgement resulting from the intentional or unintentional misuse or rating scale.

A

Rating error

73
Q

Error in rating that arises from the tendency on the part of the rater to be lenient in scoring.

A

Leniency Error

74
Q

Systematic reluctance to giving ratings at either positive or negative.

A

Central Tendency Error

75
Q

Tendency to give a particular ratee a higher rating than he of she objectively deserves because the rater’s failure to discriminate among conceptually distinct aspects of a ratee’s behavior.

A

Halo effect

76
Q

The extent to which a test is used in an impartial, just and equitable way.

A

Fairness

77
Q

Refers to a group of statistics that can be calculated for individual test items.

A

Item analysis

78
Q

What are the 2 commonly used techniques of item analysis?

A

Item Difficulty
Item Discrimination

79
Q

A commonly used technique of item analysis that is appropriate for maximal performance test achievement and aptitude test. It requires that test items be scored as correct and incorrect.

A

Item Difficulty

80
Q

Percentage of the pupils who got the item right. It can also be interpreted as how easy or how difficult an item is.

A

Difficulty Index/ Item Difficulty Index

81
Q

Formula for Item Difficulty index

A

P= number of person answering item correctly/ N (total number of people taking the test)

82
Q

What is the value for optimal item Difficulty?

A

Between .5-.7

83
Q

What does a high p value mean in item difficulty index?

A

Most people are correct on the item.

84
Q

What does a low p value mean in item difficulty index?

A

Most people answered the item incorrectly.

85
Q

Table of equivalent in interpreting the difficulty index:

_= Very difficult
.21-.80= moderately difficult
.81-1.00= _.

A

.00-.20
Very easy

86
Q

The presence of many items at the _ level limits the potential reliability and validity of the test.

A

0

87
Q

The presence of many items at _ level reduces variability and limits the test’s potential reliability and validity.

A

1

88
Q

A commonly used technique of item analysis that is appropriate for almost any type of test. It indicates the extent to which different types of people answer an item in different ways.

A

Item Discrimination

89
Q

It separates the bright test takers from the poor ones.

A

Discrimination Index

90
Q

What are the 2 approaches for measures of item Discrimination?

A

Item discrimination index
Item-total correlation

91
Q

The proportion obtained by comparing the performance of 2 subgroups of test takers, used in maximal performance testing. Also known as D.

A

Extreme Group Method