Test Development Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

is an umbrella term for all that goes into the process of creating a test

A

Test Conceptualization

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

The thought that “there ought to be a test for…” is impetus to
developing a new test.

A

TC

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

The stimulus could be knowledge of psychometric problems with other
tests, a new social phenomenon, or any number of things.

A

T C

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

The process of setting rules for
assigning numbers in measurement.

A

Scaling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

are instruments
to measure some trait, state, or ability and may be categorized in many ways

A

Scaling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

was very influential in the
development of sound scaling methods

A

LL Thorndike

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

grouping of words, statements, or symbols on which
judgments of the strength of a particular trait, attitude, or emotion are
indicated by the test taker.

A

Rating Scales

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Developed to be “a practical means of assessing
what people believe, the strength of their convictions, as well as individual differences in moral
tolerance” (p

A

Morally Debatable Behavior Scale Revision

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Each item presents the test taker with five alternative responses (sometimes seven), usually on an agree–disagree or
approve–disapprove continuum.

A

Likert Scale

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Offers a continuum of responses that allow for measurements of
attitudes on various topics

A

Likert Scale

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Test takers must choose between two alternatives according to some rule.

A

Method of Pair Comparisons

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

For each pair of options, test takers receive a higher score for
selecting the option deemed more justifiable by the majority of a group
of judges.

A

Method of Pair Comparisons

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Entails judgments of a stimulus in
comparison with every other stimulus on the scale

A

Comparative Scaling (Sorting Task)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Stimuli are placed into one of two or more
alternative categories that differ quantitatively with respect to some
continuum.

A

Categorical Scaling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Items range sequentially from weaker to stronger
expressions of the attitude, belief, or feeling being measured.

A

Guttman Scale

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

provide a list of terms
and the individual
selects that most
characteristic of
herself or himse

A

Adjective checklist

17
Q

provide a list of adjectives that must be sorted into nine piles of increasing similarity to the target person.

A

Q - Sorts

18
Q

Guide for item Writing

A

Define clearly what you wish to measure
2. Generate pool of items
3. Avoid items that are exceptionally long
4. Be aware of the reading level of those taking the scale and the
reading level of the items
5. Avoid items that convey two or more ideas at the same time
6. Consider using questions that mix positive and negative
wording

19
Q

The reservoir or well from which items will or will not be
drawn for the final version of the tes

A

Test Pool

20
Q

Includes variables such as the form, plan, structure, arrangement, and layout of individual test items.

A

Item Format

21
Q

Items require test takers to select a
response from a set of alternative responses

A

Selected-response format

22
Q

Items require test takers to supply or to create the correct answer, not merely to select it.

A

Constructed response format

23
Q

Multiple-choice format has three elements:

A

1) a stem, (2) a correct
alternative or option, and (3) several incorrect alternatives or options
variously referred to as distractors or foils.

24
Q

Distractions

A

b: standardized behavioral
samples; c: reliable assessment instruments; and d: theory-linked measures

25
Q

A relatively large and easily accessible collection of test
questions.

A

Item Bank

26
Q

An interactive, computer-
administered test-taking process wherein items presented to the test
taker are based in part on the test taker’s performance on previous
items.

A

Computerized Adaptive Testing

27
Q

A discrepancy between scoring in an anchor protocol and the scoring
of another protocol is referred

A

Scoring Drift

28
Q

refers to the revalidation of a test on a sample of
test takers other than those on whom test performance was originally
found to be a valid predictor of some criterion.

A

Cross Validation

29
Q

test validation process conducted on two or more
tests using the same sample of test takers.

A

Co validation

30
Q

Allows test developers to evaluate the validity of
items in relation to a criterion measure.

A

Item validity Index

31
Q

Indicates how adequately an item separates
or discriminates between high scorers and low scorers on an entire test.

A

Item discriminatory Index

32
Q

The quality of each alternative within a
multiple-choice item can be readily assessed with reference to the
comparative performance of upper and lower scorers.

A

Analysis of item alternatives:

33
Q

is an item that favors one particular group of
examinees in relation to another when differences in group ability are
controlle

A

Biased Test Item