Evidence-based evaluation Flashcards
What kind of evaluations are there to assess the strategy’s effectiveness?
- EMMIE Framwork
- The Maryland Scale (SMS)
EMMIE rates each intervention against 5 dimensions. What are these dimensions? And what do they do?
- Effect –> what is the impact on the crime.
- Mechanisms/mediation –> how does it work; What is it about the intervention that could explain its effect.
- Moderators –> where does it work; what are the circumstances and contexts where the intervention is likely to work or not work.
- Implementation –> how to do it; the conditions that should be considered when implementing an intervention.
- Economic-costs –> how much does it cost; direct and indirect.
What are the 4 criteria of methodological quality? And what does it mean?
- Statistical conclusion validity –> the cause precedes the effect.
- Internal validity –> the cause is related to the effect.
- Construct validity –> there is no other explanation for the effect.
- External validity –> the results can be applied to different people/places/times.
What are the main threats to the statistical conclusion validity?
- Low statistical power (e.g., due to sample size)
- Using inappropriate statistical methods
What are threats to the internal validity?
- History: the effect is caused by some event occurring at the same time as the intervention.
- Maturation: the effect reflects a continuation of preexisting trends (normal human development).
- Instrumentation: the effect is caused by a change in the method of measuring the outcome.
- Testing: the pretest measurement causes a change in the posttest measure.
- Differential attrition: the effect is caused by differential loss of units from experimental compared to control conditions.
What are the main threats to construct validity?
- Whether the intervention successfully changed what it was supposed to change (e.g., implementation issues, treatment fidelity)
- The accuracy and reliability of the outcome measure (e.g., whether police-recorded crime rates reflect actual crime rates).
Bad science = bad policy
What can we do?
- Adhere to scientific standards.
- Collect and evaluate studies
- Examine theoretical mechanisms
- Test, retest, repeat
The Maryland Scale (SMS) ranks research design from 1 to 5, what are these levels? And which are (in)appropriate?
Level 1: correlation, single point in time –> inappropriate
Level 2: before and after, no control group –> inappropriate
Level 3: before and after, with control group –> appropriate
Level 4: before and after, multiple control groups, and control for other variables –> appropriate
Level 5: randomized assignment to comparable groups; randomized controlled trials (RCT) –> appropriate
Which questions do you have to ask, after ranking the SMS to an intervention/evaluation?
Statistical conclusion validity:
1. Was the statistical analysis appropriate?
2. Did the study have low statistical power to detect effects because of small samples?
3. Was there a low response rate or differential attrition?
Construct validity:
4. What was the reliability and validity of measurement of the outcome?
What do you have to do if you have serious implementation problems? (SMS)
You have to deduct the level of the Maryland Scale by one point, for example when the attrition is really high.
Reviewing the collective evidence of the SMS, what works? What doesn’t work? What is promising?
What works –> at least two level 3-5 evaluations showing effects.
What doesn’t work –> at least two level 3-5 evaluations showing no effects.
What is promising –> at least one level 3-5 evaluation showing effects.
What are the main threats to external validity?
Differences in how causal relationships vary across types of people, settings, interventions, or outcomes.