Chapter 8 Principles of Measurement Flashcards

1
Q

Construct validity

A
  1. Extent to which an instrument or test measures an intended hypothetical concept or construct.
  2. The most valuable - yet the most difficult (takes years) - way to assess an instrument’s validity.
  3. Example: The Safe Sex Behavior Questionnaire was determined to have construct validity based on significant correlations with measures of risk-taking and self-expression.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Content validity

A
  1. Extent to which an instrument or test measures an intended content area. The most common type of validity.
  2. Determined by a panel of experts. Example: Experts in nutrition were chosen to assess the items in the Barriers to Healthy Eating Scale.
  3. Usually this type of validity is used in the development of a questionnaire, interview schedule, interview guide, or instrument.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Continuous variable

A
  1. Variable that takes on an infinite number of different values presented on a continuum.
  2. Examples: Age in years; education in years.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Criterion-related validity

A
  1. Extent to which an instrument or test measures a particular concept compared with a criterion.
  2. Measured by a validity coefficient; a higher coefficient indicates high criterion-related validity.
  3. Example: A new instrument is compared with an older, more reputable instrument.
  4. Two types: Concurrent and predictive.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Cronbach’s alpha (coefficient alpha)

A
  1. Widely used index of the extent to which a measuring instrument is internally stable.
  2. A measurement of the extent to which all the items in an instrument measure the same concept.
  3. An acceptable level of reliability for any measurement instrument is an alpha coefficient of 0.70.
  4. Can be used with instruments composed of items that can be scored with three or more possible values, such as a Likert-type scale.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Dichotomous variable

A

A nominal variable that consists of two categories.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Instrument

A
  1. A device, piece of equipment, or paper-and-pencil test that measures a concept or variable of interest.
  2. Examples from nursing research: Questionnaires, surveys, and rating scales.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Interval level of measurement

A
  1. Level of measurement characterized by a constant unit of measurement or equal distances between points on a scale.
  2. Possesses all characteristics of a nominal and ordinal scale in addition to having an equal interval size based on an actual unit of measurement, but lacks a true zero point.
  3. May be referred to as “interval data” or “interval variables.”
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Measurement

A
  1. The assignment of numerical values to concepts, according to well-defined rules.
  2. The numerical values reflect properties of the concepts under study.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Nominal level of measurement

A
  1. The lowest level of measurement, which consists of assigning numbers as labels for categories. These numbers have no numerical interpretation (i.e., they are not stating that one category has “more” and one category has “less”).
  2. Data simply show the frequency of subjects or objects in each category.
  3. May be referred to as “nominal scale,” “nominal data,” “nominal measurement,” or “categorical data.”
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Operational definition

A
  1. A definition that assigns meaning to a variable and the terms or procedures by which the variable is to be measured.
  2. Especially important for quantitative research. Usually found in the methods section.
  3. Concepts such as “spiritual well-being” must be translated to measurable definitions that are valid reflections of the concepts.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Ordinal level of measurement

A
  1. Level of measurement that yields rank-ordered data (i.e., highest to lowest, most to least).
  2. Specifies the order of items being measured, without specifying how far apart they are.
  3. May be referred to as “ordinal scale,” “ordinal data,” “ordinal variables,” or “ordinal measurement.”
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Psychometric evaluation

A

Evaluating properties of reliability and validity in relation to instruments being used to measure a particular concept or construct.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Predictive validity

A
  1. Ability to predict future events, behaviors, or outcomes.
  2. Example: A university admissions committee uses applicants’ GRE scores to decide whether to admit them to a graduate program, based on the idea that a high GRE score is predictive of success in their program.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Ratio level of measurement

A
  1. The highest level of measurement, characterized by equal distances between scores having an absolute zero point.
  2. The zero point indicates absolutely none of the property.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Reliability

A
  1. Value that refers to the consistency with which an instrument or test measures a particular concept. Different ways of assessing reliability include test-retest, internal consistency, and interrater.
  2. An instrument with high reliability should yield essentially the same scores each time the test is administered.
  3. Asks, “Is it consistently generating the same measurements?”.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Test-retest reliability

A
  1. An approach to reliability examining the extent to which scores are consistent over time (their stability).
  2. A researcher measures a group of individuals twice with the same measuring instrument or test, with the two testings separated by a particular period of time.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Validity

A
  1. Value that refers to the accuracy with which an instrument or test measures what it is supposed to measure. Different types of validity include content, criterion, and construct.
  2. Asks, “Are the measurements meaningful?”.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Relationship between reliability and validity

A
  1. Both are essential to the meaning and accuracy of the results produced by an instrument.
  2. They are the two most important issues to consider when examining the worth of any instrument used to measure variables in a study.
  3. Reliability refers to whether an instrument generates consistent measurements; validity refers to whether an instrument measures what it is supposed to measure.
  4. An instrument’s data must be reliable if they are to be valid; an instrument can be reliable without being valid.
  5. They are context/population-specific - an instrument designed for premenopausal women is not reliable and valid for homeless youth, unless tested and proven.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

In order to make sense out of data collected, each variable must be _

A

Defined operationally using a variety of measurement techniques.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

The selection of a measurement technique depends upon _

A

The particular research question and the availability of instruments.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

The four levels of measurement

A
  • In order of increasing sophistication*:
    1. Nominal (lowest level - numbers are merely labels for categories).
    2. Ordinal (more precise than nominal, but does not lend itself well to statistical analysis).
    3. Interval (equal interval sizes and units allow addition, subtraction, and computation of averages, but there is no true zero point).
    4. Ratio (highest level - has an absolute zero point; data can be manipulated using any arithmetic operation).
23
Q

Examples of nominal data

A

The numbers are labels arranged in random order, and do not mean that one category is “more” or “less” than another:

  1. Gender (1 for males, 2 for females).
  2. Religion (1 for Catholic, 2 for Protestant, 3 for Jewish).
  3. Blood type (1 for type AB, 2 for type A, 3 for type B, 4 for type O).
24
Q

If there are only two categories associated with nominal data (such as gender), then the variable of interest is said to be _ in nature.

A

Dichotomous.

25
Q

Many instruments and scales used to measure _ variables yield ordinal data.

A

Psychosocial. (For this reason, ordinal data are common in nursing research.)

26
Q

Examples of ordinal data

A

Only the order of categories is important, as intervals between ranks are not equal:

  1. Satisfaction (1 for “very satisfied;” 2 for “somewhat satisfied;” 3 for “not satisfied”).
  2. Education (1 for 6 years or less; 2 for 8 years; 3 for 11-12 years; 4 for 13-16 years).
  3. Household income (1 for $0-$4,999; 2 for $5,000-$9,999; 3 for $10,000-$19,999; 4 for $20,000-$29,999; 5 for $30,000-$49,999).
27
Q

Many instruments and scales used in nursing research report “scores” and are usually referred to as _

A

Interval data or interval variables. (Many highly reliable psychosocial instruments or scales yield interval data.)

28
Q

Example of interval data

A

Numerical values with equal interval sizes and an actual unit of measurement, but no meaningful zero point:
Temperature (Fahrenheit or Celsius: 0° does not indicate an “absence” of temperature; 90°F is not “twice as hot” as 45°F).

29
Q

Examples of ratio data

A

Categories are different, ordered, separated by a constant unit of measurement, and have an absolute zero point:

  1. Age (e.g., number of years).
  2. Time (e.g., number of hours).
  3. Weight (e.g., number of kilograms).
  4. Length (e.g., number of inches).
30
Q

From a statistical standpoint, _ and _ measurements are treated as the same.

A

Interval and ratio. (The mean and standard deviation can be calculated for both of these.)

31
Q

Measurement error

A
  1. How well or how poorly a particular instrument performs in a given population.
  2. Can be random (scores vary in a random way) or systematic (scores are incorrect, but they are incorrect in the same direction).
32
Q

_ measures are by far the most commonly used data collection measures in nursing research.

A

Psychosocial.

33
Q

Reliability coefficient (r)

A

A numerical expression of reliability between 0.00 and 1.00; a higher value indicates higher reliability (r = 0.00 is a completely unreliable test, and r = 1.00 is a completely reliable test).

34
Q

3 aspects considered by reliability testing and 3 corresponding measures of reliability

A
  1. Stability (the consistency of repeated measurements): Test-retest reliability.
  2. Homogeneity: Internal consistency.
  3. Equivalence: Interrater reliability.
35
Q

Test-retest reliability and time intervals

A
  1. Reliability is generally higher when the time lapse between testing is short, usually no longer than 4 weeks.
  2. Caution is advised when reading research reports regarding how stable the measurement is when researchers do not indicate the length of time between the two testings.
36
Q

Internal consistency

A
  1. Consistency across items of an instrument, with individual items being individual questions. To the extent that certain items “hang together” and measure the same thing, an instrument is said to possess high internal consistency reliability.
  2. Can be assessed by calculating Cronbach’s alpha, which should be 0.70 or higher.
37
Q

Kuder-Richardson 20 (KR-20)

A

A statistical procedure used to assess internal consistency when the items of an instrument are scored dichotomously (e.g., “yes” or “no”, “true” or “false”).

38
Q

Interrater reliability

A
  1. An approach to reliability examining the extent to which raters are in agreement on something.
  2. A researcher has raters evaluate a particular situation, then assesses the consistency among the raters.
39
Q

Cohen’s kappa

A

A statistical procedure used to assess interrater reliability for situations involving nominal data in which raters classify the items being rated according to discrete categories.

40
Q

Determining interrater reliability is very common in nursing research and is used in many _ studies.

A

Observational. (Interrater reliability values should be reported in any study in which observational data are collected or when judgments are made by two or more data collectors).

41
Q

Concurrent validity

A
  1. When a new instrument is administered at about the same time that data are collected on the criterion.
  2. Example: An instrument designed to assess well-being and an instrument designed to assess self-esteem are administered in the same sitting.
42
Q

“Stability” is associated with _

A

Test-retest reliability.

43
Q

“Cronbach’s alpha” is associated with _

A

Internal consistency.

44
Q

“Kappa statistic” is associated with _

A

Interrater reliability.

45
Q

Pilot testing

A

Means of evaluating a newly developed or adapted instrument, especially when there is no relevant previous research; identifies problems in an instrument before it is administered in its final form.

46
Q

As a nurse, you rate a patient’s pain on a scale from severe to low. This results in which type of scale?

A

Ordinal.

47
Q

Which of the following is not an assumption underlying testing and measurement? (Various approaches to measuring aspects of the same thing can be useful; Error is rarely present in the measurement process; Present day behavior predicts future behavior; Testing and assessment benefit society.)

A

Error is rarely present in the measurement process.

48
Q

The _ the level of measurement, the greater flexibility in choosing statistical procedures.

A

Higher. (The highest level of measurement is going to provide the maximum amount of information.)

49
Q

Nominal data can be either _

A

Dichotomous (e.g., gender) or categorical (e.g., marital status).

50
Q

Measurement error can be decreased by increasing the _

A

Intervention fidelity - using the same data collection methods with the same type of subjects at the same point in time.

51
Q

Type I error

A

A false positive (detecting something that was not actually there) - the researcher rejected a null hypothesis that was true. Considered more serious.

52
Q

Type II error

A
  1. A false negative (failing to detect something that was present) - the researcher accepted a null hypothesis that was false.
  2. Occurs when a sample size is not large enough (“underpowered”) - can minimize through the use of power analysis.
53
Q

Reliability is concerned with _ error and validity is concerned with _ error.

A

Random; systematic.