TM T3 Flashcards

1
Q

Validity

A

The degree to which empirical evidence and theoretical rationales support an assessment conclusion.

Not something a test HAS; a justification

Criterion, construct, content

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Construct validity

A

Based on conceptual variables underlying a test.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Content validity

A

Based on subject matter of a test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Avoid saying

A

“the validity of the test”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

When is validity wrong?

A

Wrong population -> different groups require special tools

Wrong task -> using the wrong test can lead to invalid results

Wrong context -> wrong testing environment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Wrong context example

A

Using a personality test for a hiring decision

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Face validity

A

the APPARENT soundness of a test or measure regardless if they are. Intuitive. Something can lack this but still be valid - that might even be BETTER, since you cannot tell what is being tested

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

5 Sources of Evidence of Validity

A
  1. Test content
  2. Response processes
  3. Internal structure
  4. Relations with other variables
  5. Consequences of testing
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Content validity details

A

Evidence based on a test’s content
Extent to which a test measures subject matter or behavior under investigation
Example: test for 3rd grade math… how well does it represent what we want kids to know?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Validity based on response processes

A

How do people actually respond to a test - is it measuring what it’s supposed to measure?

“Can you repeat the question in your own words?”
“What, to you, is ___”
“How sure are you of your answer?”

Revise based on answers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Evidence based on internal structure

A

Does the test align with the theoretical framework or construct it’s intended to measure? Is the construct represented in the patterns of responses that people give?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Factor analysis

A

Tool used to see which variables correlate with eachother and if they cluster around one common factor: “energy loss, appetite changes, difficulty concentrating” might cluster around somatic traits in the BDI-II

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Criterion validity

A

How well a test correlates with a specific standard
4 methods
Predictive, concurrent, retrospective and incremental
If a measure of criminal behavior is calid, we should be able to tell if:
1. They will be arrested in the future
2. The are currently breaking th elaw
3. They have a previous criminal record

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

2 ways of thinking of CONSTRUCT validity

A

Convergent - how does my test correlate to other similar tests?

Discriminant - can a test prove you’re measuring your construct and not something else?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Evidence based on consequences of testing

A

“Does this test produce situations for people that are not OK?”
If so -> not as strong as a test with no negative outcomes
Does this test promote fair outcomes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Operational definitions

A

How you’re going to try to measure something - as an external, observable behavior

Example: aggression
How many times did they hit somebody

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Construct

A

Abstract idea
Can’t directly measure it
Characteristics or attributes
EX: aggression, intelligence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Demonstrating content validity

A

Define the test universe
Develop the test specifications
Establish the test format
Construct the test questions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Construct validity: defining the test universe

A
  1. What relevant research is there to help develop the constructs?
  2. Who are the key experts? Can they evaluate items? Also - cite them
  3. Main construct aspects/dimensions
  4. Other validated instruments
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Construct validity: Test specifications, format, questions

A
  1. Specify content areas
  2. Construct test questions
  3. Purpose and intended use of test
  4. Format and length
  5. Et cetera
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

IVR example

A

Constructs: minimization, violence recognition, partner blaming, distal blaming

22
Q

IVR lacks some

A

test specifications

23
Q

Criterion-related validity

A

Correlation with establish CRITERION (standard of comparison) - how well it correlates with its established standard

24
Q

Examples of criterion validity

A

SAT performance correlating to academic success

accidents on a job correlating to supervisor ratings

25
4 types of criterion validity
Predictive Concurrent Retrospective Incremental
26
Predictive validity (criterion)
Test scores correlate with a future variable Assessed after test is admin. Prospective validity EXAMPLE: predicting disease onset
27
Concurrent validity (criterion)
Measurement backed by a related measurement at the same time - like a self-report and a supervisor rating. Convergent validity.
28
Retrospective validity
Correlates with past occurrences - accident proneness and past injuries. Postdictive validity.
29
Incremental validity
Does a test predict beyond other measures? Controlled regression for effects of other variables Example: CROPS and PROPS - assessed incremental validity of CROPS/PROPS by seeing if they added to prediction of lifetime traumatic events rating
30
Validity coefficient
+- .10 weak +- .30 moderate +- .50 strong
31
Cohen validity coefficient
+- .10 weak +- .30 moderate +- .50 strong
32
Linear regression
Y1 = a + bx Slope: expected change in y for every 1 unit change in x Intercept: where the regression line crosses the y axis Deviations from line: residuals
33
Coefficient of determination
Proportion in variability of y accounted for by variability in x - amount of shared variance between the predictor and the criterion
34
All about validity coefficient
Relationship between test score and criterion measure EX: correlation between PREP-M scores (prep for marriage) and reports of marital satisfaction.
35
Measurement error is also called
Attenuation due to unreliability
36
Threats to validity
Criterion contamination Restriction of range Attenuation due to unreliability
37
Criterion contamination
Criterion is influenced by unrelated factors - sales influenced by territory
38
Restriction of range
Limited variability in predictor or criterion (SAT scores in highly selective colleges may not predict as well for academic performance due to low diversity in range)
39
Attenuation due to Unreliability
measurement error
40
Think of the criterion variable as
the dependent variable, or the existing standard used for a comparison
41
Reliability vs. validity
R: consistency V: accuracy
42
True score
Hypothetical score without any error - can't be directly observed or measured
43
Observed score
Actual score, probably with some degree of measurement error
44
Systematic error
"the scale that always adds 5 pounds or the scale that always subtracts 5 pounds" Consistent error in measurement that skews the result in one direction Impacts validity EXAMPLE: observer bias, respondent bias
45
Random error
Unpredictable Impacts reliability EX: mood fluctuations, guessing on MCs
46
Internal Consistency reliability
the extent to which test items are measuring the same construct alpha EX: on an anxiety questionnaire, someone with anxiety who rates high on feeling worried is likely to rate high as nervous on another item
47
Alternate-forms reliability
consistency of results from multiple tests designed to measure the same subject
48
Interrater reliability
degree of agreement between raters rating consistency
49
What is the difference between heterogenous tests and homogenous tests?
Heterogenous test: measuring broader construct Example: WAIS measuring verbal comprehension, memory, processing speed) or the SAT and its subjects; BDI - depression as a broad subject Homogenous test: Measuring a single construct Example: basic math test
50
*Why is it necessary to include confidence intervals to interpret individual test scores?
they offer the individual's true score, quantifying possible error as a warning
51
Longer tests...
Have more items, increasing chances of reliability
52