Variables and Their Measurement Flashcards

1
Q

Variable (definition)

A

Characteristic of an individual, object, or environmental conditoin

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Types Of Variables (3)

A

1) Independent Variable
2) Dependent Variable
3) Extraneous Variable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Independent Variable (def)

A

The variable that is intentionally manipulated by researchers to produce a change in an outcome; there are levels indicated by the number of IV’s in the study

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

IV’s are used in ___ Types of studies (2)

A
  • Interventional

- Prognostic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

IV’s are NOT used in ___ Studies (1)

A

Descriptive; bc researchers don’t manipulate anything and there is no outcome of interest

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Purpose of IV’s in Interventional Studies (2)

A
  • Can draw CAUSE AND EFFECT relationships because manipulating an IV
  • They are commonly used as treatments/interventions
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Levels of IV’s (3)

A
  • > 1 IV = factorial design
  • 2x2 factorial design
  • 2x3 factorial design
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

2x2 factorial design?

A

Means there are 2 IV’s with 2 levels

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

2x3 factorial design?

A

Means there are 2 IV’s with 3 levels

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

IV Prognostic Studies (3)

A
  • CANNOT draw Cause and Effect relationships bc not manipulating anything!
  • IV’s are used to predict the outcome of interest
  • Not intentionally manipulated by researchers
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Dependent Variable (def)

A

The outcome of interest

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Purpose of IV’s and DV’s in Interventional Studies

A

-Manipulate IV to cause a change in the DV= CAUSAL relationship!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

IV’s and DV’s in Prognostic Studies

A

-Assess whether the IV is associated with the DV or predicts the DV; looks at IV and tries to predict the DV

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Extraneous Variable(def)

A
  • An outside variable that influences the DV, but it’s not an IV- want to limit these!
  • Should be controlled through study design and statistical adjustment so you can say without a doubt that change was from your tx
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Discrete Variable (def)

A

Values are distinct categories

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Dichotomous Variable (def)

A
  • A type of discrete variable

- Variables who only have 2 possible values; ex. Male and Female

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Continous Variable (def)

A

Variables who have a theoretical infinite number of values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Levels of Measurement (4)

A

1) Nominal
2) Ordinal
3) Interval
4) Ratio

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Nominal Level of Measurement

A
  • Includes values that are named CATEGORIES
  • NO rank; no one category is better than another
  • Statistical analysis uses FREQUENCIES (not means)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Ordinal Level of Measurement

A
  • Categories have a RANK order relative to one another; some form of hierarchy (ex. Likert scales)
  • Statistical Analysis uses FREQUENCIES (not means)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Interval Level of Measurement

A
  • Assigns quantitative values to variables
  • Do NOT have a KNOWN ZERO POINT, but zero doesn’t mean there is an “absence” of the characteristic
  • Possible values extend to infinity and can be both positive or negative
  • Values may be added and subtracted, but not multiplied or divided
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Ratio Level of Measurement

A
  • Have a RANK order, and a KNOWN ZERO POINT
  • No negative values
  • has a TRUE zero
  • Values can be added, subtracted, multiplied, and divided
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

True zero (def)

A

If there is a zero present, there is an ABSENCE of the characteristic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What type of variable is MMT grades? Why?

A

Discrete and ordinal variable; not an infinite number of possibilities and there is a heirarchy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

What type of variable is gender? Why?

A

Discrete (dichotomist) and Nominal variable; bc not an infinite number of possibilities and equal groups (no hierarchy)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

What type of variable is Blood Pressure? Why?

A

Ratio Variable- bc there is a true zero (no negative numbers)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

Reliability (def)

A

How consistent and free from error is the instrument; represents how much of a measurement represents error and how much is the true score

28
Q

Measurement error (def)

A

Difference between observed and true score

29
Q

Observed score =

A

True score-error

30
Q

How researchers minimize measurment error?

A

Choose instruments which have established reliability measures

31
Q

2 forms of Measurement Relaibility

A
  • related to the instrument

- Related to the person TAKING the measurements

32
Q

Types of Reliability related to the INSTRUMENT (2)

A

1) Reproducibility aka Test-Retest Reliability

2) Internal Consistency

33
Q

Reproducibility (def)

A
  • Aka Test-Retest Reliability

- established when an instrument is used in two separate occasions with the same subjects

34
Q

Internal Consistency (def)

A

Related to self-report outcome measure

35
Q

Types of Reliability related to the RATER (2)

A

1) INTRA-rater reliability

2) INTER-rater Reliability

36
Q

Intra-rater Reliability (def)

A

Consistency of repeated measures performed by ONE INDIVIDUAL

37
Q

Inter-rater Reliability (def)

A

Consistency of measures performed by >1 individual

38
Q

Regression Toward the Mean

A

Observed scores move closer to the mean with repeated tests; affects OUTLIERS

39
Q

Reliability Coefficient(def)

A

Estimates reliability based on statisical concept of variance; a Measure of variability or differences among scores within a sample

40
Q

Reliability Coefficient (math)

A
  • (True score variance/(true score variance + error variance))
  • Ranges from 0.00-1.00
41
Q

Reliability Coefficient that indicates POOR reliability:

A
42
Q

Reliability Coefficient that indicates MODERATE reliability

A

.50-.75

43
Q

Reliability Coefficient that indicates GOOD Reliability

A

> .75

44
Q

Validity

A

Does the test measure what it’s supposed to measure; goes hand-in-hand with reliability- don’t want to be reliably invalid!

45
Q

4 Types of Validity

A

1) Face Validity
2) Content Validity
3) Construct Validity
4) Criterion Validity

46
Q

Face Validity (def)

A
  • Simplest and most subjective form of validity

- The instrument appears to be the appropriate choice to measure the variable (or not)

47
Q

Content Validity (def)

A

Does the measure represent all of the relevant facets of the variable of interest?; should not contain elements that capture irrelevant information

48
Q

Construct Validity (def)

A

Does the measure reflect the operational definition of the concept or construct it says it represents?

49
Q

Convergent Validity (def)

A

Method used to evaluate the construct validity of an instrument; assess the relationship between scores on the instrument of interest and on another instrument ; if scores from both instruments yield similar resulst, then the measure has convergent validity

50
Q

Discriminant Validity (def)

A

Method used to validate construct validity of an instrument; reflects degree to which an instrument can distinguish between different constructs

51
Q

2 Things that evaluates Construct Validity

A

1) Convergent Validity

2) Discriminant Validity

52
Q

Convergent Validity (def)

A
  • Method used to evalute construct validity of an instrument;
  • Assess the relationships between scores on the instrument of interest and ON ANOTHER INSTRUMENT
  • If scores from both instruments yield similar results, then the measure has convergent validity
53
Q

Discriminant Validity (def)

A
  • Method used to validate construct validity of an instrument
  • Reflects the degree to which an instrument can distinguish between different constructs; ex. can a goniometer distinguish between strength and ROM?
54
Q

Things that Evaluate Criterion Validity (2)

A

1) Concurrent Validity

2) Predictive Validity

55
Q

Criterion Validity (def)

A

-Degree to which the scores on an instrument are related to the scores on a reference standard instrument

56
Q

Concurrent Validity (def)

A

-Method of evaluating criterion validity; administer the test of interest and reference standard test (the GOLD STANDARD) at the same time (to eliminate bias)

57
Q

Predictive Validity (def)

A
  • Method of evaluating criterion validity

- Degree to which the results from the test of interest can predict a future outcome

58
Q

Sensitivity and Specificity are modes of ____ Validity

A

Criterion

59
Q

A test CANNOT be ___ but not ___

A

valid;reliable

60
Q

Instruments should be ____, ____, and ____ _ ____

A

reliable, valid, responsive to change

61
Q

To be Responsive (to change) an instrument should Have: (2)

A
  • Construct validity

- Severeal values on the scale to limit the floor or cieling effect

62
Q

Standard error of Measurement (def)

A

Extent to which observed scores are disbursed around the “true” score

63
Q

Large Standard of Error = Less ___; why?

A

Responsive; bc the “true” values are lost in the inaccuracies each time the measure is repeated

64
Q

Ceiling effect

A

Happens when scores don’t go high enough

65
Q

Floor Effect

A

Happens when scores don’t go low enough