Exam 2 Objectives Flashcards
Distinguish among continuous, discrete, and dichotomous variables.
Continuous variables: can theoretically have any value along a continuum within a defined range.
- Ex: weight in pounds
Discrete variables: can only be described in whole integer units.
- Ex: HR in BPM
Dichotomous variable: can only take on two values.
- Ex: yes or no on a survey
Discuss the challenge of measuring constructs.
Measuring constructs are subjective, abstract variables.
They are measured according to expectations of how a person who possesses the specified trait would behave, look, or feel in certain situations.
Only reflects something w/ in a person, does not exist as an externally observable event (latent trait)
Define and provide examples of the four scales of measurement.
Nominal: category/classifications
- Ex: blood type, gender, dx
Ordinal: numbers in rank order, inconsistent/unknown intervals. Based on greater than, less than
- Ex: MMT, function, pain
Interval: numbers have rank orders and equal intervals, but no true zero. Can be added or subtracted, cannot be used to interpret actual quantities.
- Ex: Fahrenheit vs Celsius, shoe size
Ratio: numbers represent units w/ equal intervals measured from true zero
- Ex: Height, weight, age
Discuss the relevance of identifying measurement scales for statistical analysis.
Determination of which mathematical operations are appropriate.
Determination of which interpretations are meaningful.
Statistical procedures
- Parametric tests: apply mathematical manipulations, requiring interval or ratio data… Mean, median, mode.
- Nonparametric tests: do not make the same assumptions and are designed to be used with ordinal and nominal data.
Discuss the importance of reliability in clinical measurement.
Relative reliability: Reflects true variance as a proportion of total variance in a set of scores.
Intraclass correlation coefficients (ICC) and kappa coefficients are commonly used.
Absolute reliability: Indicates how much of a measured value, expressed in the original units, is likely due to error.
Most commonly uses standard error of measurement.
Define reliability in terms of measurement error.
Classical measurement theory: any observed score involves true score (fixed value) and unknown error component (small or large)
Difference between true score and observed value = measurement error
Observed score = true score ± error component
Distinguish between random and systematic error.
Random errors are a matter of chance, possibly arising from factors such as examiner or subject inattention, instrument imprecision, or unanticipated environmental fluctuation.
- Can occur through the measuring instrument itself for example: Imprecise instruments or environmental changes affecting instrument performance can also contribute to random error.
Systematic errors: are predictable errors of measurement. They occur in one direction, constantly overestimating or underestimating the true score.
- Systematic errors are consistent. Consequently, systematic errors are not a threat to reliability. Instead, systematic errors only threaten the validity of a measure.
Describe typical sources of measurement error.
The person taking the measurements — the raters.
The measuring instrument.
Variability in the characteristic being measured.
Describe the effect of regression toward the mean in repeated measurement.
Regression toward the mean (RTM) = statistical phenomena when extreme scores are used in the calculation of measured change. Extreme scores on an initial test are expected to move closer (or regress) toward the group average (mean) on a second test.
RTM, if not considered, can interfere when researchers try to extrapolate results observed in a small sample to a larger population of interest.
Discuss how concepts of agreement and correlation relate to reliability.
Through interrater agreement and correlation, a study is likely to be more reliable.
Define and provide examples of test-retest, rater, and alternate forms reliability.
Test-retest: determines the ability of an instrument to measure subject performance consistently.
Test-retest intervals: time intervals between tests must be considered.
Carryover: carryover influenced by practice or learning during the initial trial alters performance on subsequent trails.
Testing effects: when the test itself is responsible for observed changes in a measured variable.
Rater reliability: training and standardization may be necessary for rater(s); the instrument and the response variable are assumed to be stable so that any differences between scores on repeated tests can be attributed solely to rater error.
Intrarater: stability of data recorded by one tester across two or more trials.
Interrater: two or more raters who measure same subject
Alternate forms of reliability: also called equivalent or parallel forms; assesses the differences between scores to determine whether they agree; used as an alternative to test-retest reliability when the intention is to derive comparable versions of a test to minimize the threat posed when subjects recall their responses.
Discuss how generalizability theory influences the interpretation of reliability.
Reliability exists in a context: Relevant to a tool’s application.
Reliability is not all-or-none: Exists to some extent in any instrument
Discuss how reliability is related to the concept of minimal detectable difference.
Minimal detectable change (MDC): the amount of change that goes beyond error.
GREATER RELIABILITY = SMALLER THE MDC
MDC is based on the standard error of the measurement (SEM)
SEM: The most used reliability index, it provides a range of scores within which the true score for a given test is likely to lie.
MDC is also known as minimal detectable difference, smallest real difference, smallest detectable change, coefficient of repeatability, or the reliability change index.
Discuss the importance of validity in clinical measurement.
Validity relates to the confidence we have that our measurement tools are giving us accurate information about a relevant construct so that we can apply results in a meaningful way.
Used to measure progress toward goals and outcomes.
Needs to be capable of…
- discriminating among individuals w/ and w/o certain traits, dx, or conditions
- evaluating magnitude/quality of variable
- creating accurate predictions of patient future
Define and provide examples of face, content, criterion-related, and construct validity.
Face validity: Implies that an instrument appears to test what is intended to test.
- Judgment by the users of a test after the test is developed.
Content validity: establishing multiple items make up a sample, or that scale adequately samples the universe of the content that defines the construct being measured.
- The items must adequately represent the full scope of the construct being studied.
- The number of items that address each component should reflect the relative importance of that component.
- The test should not contain irrelevant items.
Criterion-related validity: Comparison of the results of a test to an external criterion
- Index test AND Gold or reference standard as the criterion
- Two types:
* Concurrent validity (test correlates w/ reference standard at same time)
* Predictive validity
Construct validity: Reflects the ability of an instrument to measure the theoretical dimensions of a construct; establishes the correspondence b/w a target test and a reference or gold standard measure of the same construct.
- Assessing presence of a latent trait
- Methods of construct validation
* Known-groups method, Convergence and divergence, Factor analysis.
* Convergent validity: extent to which a test correlates w/ other tests of closely related structure
* Discriminant validity: extent to which a test is uncorrelated w/ tests of distinct or contrasting constructs.