2nd exam Flashcards
Validity looks at what
Accuracy, is the test measuring what it is intended to measure
The overarching validity all others fall under this one
Construct Validity
Does validity look at entire test or item quality?
Falls under reliability looking at the entire test
What are the validities under construct validity
Content, convergent, discriminant, criterion related, incremental, ecological
Construct validity
assessing the accuracy of the test to measure certain personality or psychological constructs
One problem with construct validity
is that the test is robust enough that it can accurately measure psychological constructs that often might not be stable
With construct validity, we need to understand when the theories are changing
we know the testing is changing too
To truly measure what a test is intended to measure?
it has to have a certain level of timelessness
Content validity
items need to cover the material that the instrument is supposed to cover, how relevant are the items to the construct
Content validity is looking at 2 questions
does the test cover a representative sample of specified skills and knowledge? Is the test performance reasonably free from influence of irrelevant variables?
What does content validity assume?
assumes that you have a good detailed description of the content domain- which is not always possible
What are the issues with experts administering tests with content validity?
Issue with experts is that they may be too invested or biased to the construct, or don’t have lived experienced, they only have theoretical knowledge
What happens with content validity, if measure does not have appropriate content?
you will make incorrect or erroneous clinical judgments based on measure
Why is content validity important to consider
Since development of measure maybe valid only under circumstances, different time periods, such as issues with gender nowadays
ex: cultural variability- not all cultures experience the same depression
Differences between trait-based depression and state-based depression?
Trait- endogenous- no precipitating factors, no situation caused it its chemically happened
State- theres a stressor, has precipitating factors
One way we come up with content validity is by using
Focus groups
Focus groups
allow you to go to individuals who have experienced the construct for them to help give appropriate items for that construct in a group setting
How is focus groups beneficial?
get a deeper understanding of the construct and people may feel comfortable to share their experiences in a group
Lived experiences within focus groups helps with
the accuracy of the construct, creates better validity
Focus groups were not used as much in the 1967-2002 when tests were created why?
It is hard to find a sample and it is hard to get funding for it
What are some drawbacks of content validity focus groups?
some are limited in their ability to participate (chronically mentally ill), hard to develop focus groups for rare constructs and some facilitators in focus groups will lead the group in a biased fashion
Examples of questions for focus groups
-“what was it like when you were deployed?”- for veterans to help with the language content in the test
What happens in content validity after the items have been generated through experts and focus groups?
experts will evaluate your scales and response options, help you write more clear questions
Criterion related validity
assess the degree to which the scores on an instrument accurately compare with a relevant criterion variable (a real life variable)
Criterion
real world implication
Two types of criterion related validity
Predictive and concurrent validity, both have a validity coefficient, the closer it is to 1, the more accurate it is in predicting
Predictive validity
predict an outcome (criterion) (ex: SAT predicts 1st year college GPA (real life outcome)
In predictive validity, validity coefficients look at
correlating the tests scores and the criterion variable for each person
Concurrent validity
when new test is administered and criterion variable is also collected at the same time - want to see if the measurement is matched to the criterion (real life variable)
Example of concurrent validity
(ex: chicago school interview day, we wrote a paper and had a interview (this is the real life because interaction) want to see if they are correlated
Convergent validity
testing measurement is shown to be correlated with another measurement that is examining the same construct (BDI and Hamiliton Depression Inventory should have high convergent validity)
2nd type of convergent validity
measuring the same construct but have multiple of different measurement methods (ex: self report of impulse control and family reports of impulse control problems)
convergent validity should be close to
1, which indicates that both measures are considered to be measuring the same construct
Discriminant validity
is where the testing measurement is shown to be minimally correlated with another measurement that is examining a construct that is dissimilar
With discriminant validity we want the validity to be
Close to 0 to show there is no overlap between measures or minimal association
The problem with discriminant validity
very hard to find constructs that are opposites of one another, many overlap
Face validity (not a real validity)
pertains to if the test looks valid to the examinees taking the exam, subjective and imprecise validity
Procedures to ensure validity
contrasted groups, previously available tests, criterion established by the rater, age differentiation, physical evidence of the behavior, real time observations in the world, controlled stimuli that depict variations in behaviors
Contrasted groups
give a test to two different kinds of samples and both samples have difference with regard to a specific trait, they contrast
Previously available tests
using previous tests as the criterion to compare new test
ex: WAIS always compared to WISC should be correlated with one another
Age differentiation
want to make sure the test is checked against chronological age or developmental milestones to determine whether scores increase with advancing age (ex: a 14 year old cannot do what a 17 year old can do)
Criterion established by the rater
Wants clinical interview/diagnosis done to match the instrument that was given. What you find in the instrument, should be also found in a clinical intake interview
Criterion contamination
the rater doing the clinical intake, also knows what the testing has shown, it influences their diagnoses
The criterion established by the rater can sometimes cause
criterion contamination, this is a problem with ensuring validity
physical evidence of the behavior
have people wear pedometers and check to see if reporting walking behavior (exercise)
real time observations in the world
very labor intensive (ex: GRE and college performance)
Controlled stimuli can be created that depict variations behaviors
videotape marital relationships and then see if corresponds to their depiction of self report measures of marital satisfaction
Cautions when interpreting validity
- testing procedures are not always reproduced accurately, 2. the criterion variable means nothing unless it is important or reliable, 3. make sure the population is representative of the sample, 4. need adequate sample size, 5. dont confuse criterion with predictor (when someone does well on SAT does not mean they are getting a 4.0 in college), 6. check for restricted range on both predictor and criterion, want the full normal curve, 7. is it generalizable, 8. consider differential predictions (its not that the SAT predicts GPA could be motivation
sample size does what to validity
it increases it
outliers do what to validity
can increase validity, because it is the full normal curve, you get all of the data
Incremental validity
a statistical method that measures how much more predictive a new assessment is than existing ones
Incremental validity is often seen in a chart of a
Regression Table looking at R squared
R squared means
amount of variance that can be predicted by a test
What happens when all of the R squared in a regression table are added up
Increases incremental validity
Ecological Validity is related to what type of psychology?
Neuropsychology
Can your measure be valid and not reliable?
No, if you have accuracy you also have consistency