Chapter 4 Part 1: Reliability, Validity, and Utility Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

The agreement between a test score or measure and the quality it is believed to measure.

A

Validity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

This refers to the degree to which an assessment tool produces consistent results when repeated measurement are made.

A

Consistency.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

True or False. A weighing scale always measures a person to weigh 67kg when they are in fact 60kg.

The weighing scale is consistent but not valid.

A

True.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

The four (4) different aspects of validity.

A
  1. Face validity.
  2. Content-related evidence for validity.
  3. Criterion-related evidence for validity.
  4. Construct-related evidence for validity.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

This is not really validity at all because it does not offer evidence to support conclusions drawn from test scores.

A

Face validity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

This considers the adequacy of representation of the conceptual domain the test is designed to cover.

A

Content-related evidence for validity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

True or False: You are taking an examination with coverage on the first 6 chapters of Theories of Personality (example). Suddenly, a test item about a chapter not yet discussed in class showed up.

The test no longer has content-related evidence for validity.

A

True.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

This tells how well a test corresponds with a particular criterion, and such evidence is provided by high correlations between a test and a well-defined criterion or measure.

A

Criterion-related evidence for validity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

The two (2) types of criterion-related evidence for validity.

A
  1. Concurrent validity
  2. Predictive validity.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

This refers to the forecasting function of tests.

For example, the NMAT might predict how well a high school student will perform in college.

A

Predictive validity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

In this kind of validity, the measures and criterion measures are taken at the same time.

For example, a student with a learning disability might take a diagnostic test and their school tests in the same day to determine the relationship between their learning disability and performance in school.

A

Concurrent validity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

This is established through a series of activities in which a researcher simultaneously defines some construct and develops the instrumentation to measure it.

A

Construct-related evidence for validity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

This involves assembling evidence about what a test means by showing the relationship between the test and other tests and measures.

A

Construct validation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

The two (2) types of construct-related evidence for validity are?

A
  1. Convergence evidence.
  2. Discriminant evidence.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

This refers to when a measure correlates well with other tests believed to measure the same construct.

A

Convergence evidence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

This refers to when a test should have low correlations with measures of unrelated constructs, or evidence for what the test does not measure.

A

Discriminant evidence / divergent validation.

17
Q

Six (6) guidelines for item development / item writing.

A
  1. Define clearly what you want to measure.
  2. Generate an item pool.
  3. Avoid exceptionally long items.
  4. Keep the level of reading difficulty appropriate for your intended test-takers.
  5. Avoid “double-barreled” items.
  6. Consider mixing positively and negatively worded items.
18
Q

A term for a test item that conveys two (2) or more ideas at the same time.

A

Double-barreled item.

19
Q

An item format that offers two (2) alternatives for each item.

A

Dichotomous format.

20
Q

Resembles the dichotomous format except that each item has more than two (2) alternatives.

A

Polytomous format.

21
Q

A popular format for attitude and personality scales that requires a respondent to indicate the degree of agreement with a particular attitudinal question.

A

Likert format.

22
Q

Similar to the Likert format, but has even more choices.

A

Category format.

23
Q

A format common in personality measurement in which a subject receives a long list of adjectives and indicates whether each one is characteristic of himself/herself or someone else.

A

Adjective checklist (checklists and q-sorts).

24
Q

In this format, a subject is given statements and asked to sort them into nine (9) piles. This can be used to describe oneself or to provide ratings of others.

A

Q-sort.

25
Q

A general term for a set of methods used to evaluate test items.

A

Item analysis.