Experimenal Validity & Measurement Flashcards
Review of last class:
Internal Validity
External Validity
Probabilistic Knowledge
Maturation
Testing Effects
Statistical Regression to the mean
Selection of participants
Hawthorn Effect
What is criterion referenced? (2)
Individual’s performance compared to some absolute level of performance set by a researcher
E.g. set a minimum score, set performance expectations, have age of acquisition expectation
What is the difference between Validity and Reliability?
Reliability refers to the consistency of a measure (whether the results can be reproduced under the same conditions).
Validity refers to the accuracy of a measure (whether the results really do represent what they are supposed to measure)
What is norm referenced? (2)
Individual performance compared to grp norms
Standardized when raw score standardized in some way, smoothed out to ‘normal curve’ (e.g., z, T, percentiles)
What is Measurement Reliability?
Stability, consistency of the measurement
What are the four types of test reliability?
Intrasubject reliability
Intrarater reliability
Interrater reliability
Test Retest reliablity
What is Intrarater reliability?
Consistency of the data recorded by one rater over several trials
What is Intrasubject reliability?
The reproducibility of the identical responses (answers) to a variety of stimuli (items) by a single subject in two or more trials.
What is interrater reliability?
The degree of agreement among independent observers who rate, code, or assess the same phenomenon.
Reliability is reported as (3)
correlation, standard error of measurement (SEM) or % agreement
What is validity? (2)
The degree to which an instrument measures what it is intended to measure
How appropriate, meaningful and useful are inferenced drawn from the measure
What does high validity imply? (3)
A measurement is relatively free from error.
A valid test is also reliable.
A reliable measure is not necessarily valid.
What are the 3 types of Validity?
Construct validity
Content validity
Criterion-related validity
What is construct validity? (3)
Construct validity refers to whether a test measures an abstract construct adequately. An example is a measurement of the human brain, such as intelligence, level of emotion, proficiency or ability.
What are the methods for establishing construct validity? (2)
Known Groups Method
Factor analysis
Define known groups method: (2)
- One of the methods of establishing construct validity
- Degree to which an instrument can demonstrate different scores for groups known to vary on the variable being measured.
e.g., An intelligence test should differentiate individuals with DS and TD
e.g., A measure of functional independence
Should decrease with increasing age in seniors
Should be related to the level of care needed
Should be related to severity of impairment
e.g., A measure of functional hearing
Should decrease with hearing level (dB)
Should decrease in older adults
Define Factor Analysis:
A construct is made up of a variety of dimensions
- Each dimension can be assessed using a variety of tasks
- If we measure a series of variables/items associated with a construct
Some variables would be highly correlated
Some would have little correlation
Performance on variables that are similar would cluster together
The array of clusters define the construct
What is Content Validity?
Refers to the extent to which a measure represents all facets of a given construct.
What do we mean that content validity ‘‘Indicates how adequately/fully an instrument samples the variable being measured.’’? (2)
Samples all aspects of the construct
Reflects the relative importance of each part
What do we mean by ‘‘Establishing content validity is essentially a subjective process
‘’? (3)
- No statistic measures content validity
- Content validity is established through expert opinion, review of literature, operational definitions of the test variables
- Specific to the stated objectives
Explain Face Validity: (3)
a weak type of content validity
A subjective assessment that an instrument appears to test what it is supposed to test and usually not quantified
Attempt to quantify: # of raters who assess the test as having face validity
What is Criterion-Related validity?
Extent to which one measure is related to other measures or outcomes.
Explain Criterion-Related Validity: (4)
- Most objective measure of validity
- Often assessed by correlating performance on the measure of interest and the criterion measure
- Only useful if the criterion measure is stable and valid
- The target and the criterion measurements must be measured independently and without bias (e.g. blinded)
What are the two types of criterion-related validity?
Concurrent Validity
Predictive Validity
Describe Concurrent Validity: (2)
- One of the two types of Criterion-Related Validity
- Two measures are collected at relatively the same time and the performance on each is related
e.g., # words produced in a spontaneous speech sample and PPVT-V scores
Describe Predictive Validity: (2)
- One of the two types of Criterion-Related Validity
- The measure of interest is collected earlier in time and related to a criterion measure collected later
e.g., does a parent report measure of vocabulary at 2 years predict PPVT-V scores at 4 years
A test with good validity must be able to show two things; What are they?
Convergent validity
Discriminate validity
What is Divergent validity?
Low correlation with tests that measure different constructs
What is Convergent validity?
High correlations with tests that measure the same construct
What are threats to Measurement Reliability and Validity?
Ambiguous, unclear, inconsistent instructions
Observer Bias
Reactivity
Floor and ceiling effects (Validity only)
How can Observer bias affects reliability and validity? (2)
Confirmatory bias e.g., Identification of hypernasality before/after pharyngeal flap surgery in cleft-palate patient
Carryover effects e.g., first scoring affects the second scoring
How can Observer bias reliability and validity? (2)
Confirmatory bias e.g., Identification of hypernasality before/after pharyngeal flap surgery in cleft-palate patient
Carryover effects e.g., first scoring affects the second scoring
How can reactivity affect validity and reliability?
Influences that distort the measurement e.g., participant’s awareness of measurement.
How can floor and ceiling effects affect validity? (2)
Decreases the variability of a measure
Particularly difficult when measuring change over time
What is sampling? (4)
When a Sample is drawn from a target population
Practical difficulties in accessing a target population
Therefore, we select from an accessible population
Voluntary Participation
In order to generalize a target population, sample must be: (2)
Representative: same relevant characteristics
In the same proportions
What is sampling bias?
Introduced when certain characteristics in the sample are over- or under-represented relative to the target population
What is sampling bias?
Introduced when certain characteristics in the sample are over- or under-represented relative to the target population
Explain conscious sampling bias: (2)
- Purposeful selection
- Strategic limiting of population of interest
e.g., Election poll: Voters
e.g., High-functioning autism - Limits generalization in predictable ways
- Essentially saying population of interest is a subset of total population - only generalize to the subset
Explain uconscious sampling bias: (3)
- It is a problem
- It is Unplanned and unpredictable
e.g., Election poll: how reach - land lines, cell phone, internet
e.g., People who respond to polls - Limits generalization in unpredictable ways