Evaluation design I Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

stages of evaluation

A

formative

process

outcome

The main purpose of evaluation design is to be as confidence as poss that any observed changes were cause by intervention, rather than by chance/other unknown factors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

formative eval

A

Before intervention

Acceptability and feasibility of intervention

Mainly qual, e.g. focus groups, in depth interviews

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

formative eval research

A

Penn et al. (2018)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Penn et al. (2018)

A

NHS diabetes prevention programme

eval - qual research

behav interventions

specification reflected evidence - framework for service provision

provides ev based behav intervention for prevention of T2D in high risk adults

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

process eval

A

Measures how intervention delivered and received

Mixed qual and quan

Done along the way - make alterations if necessary

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

process eval research

A

Sanchez et al. (2017)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Sanchez et al. (2017)

A

improve understanding of underlying mechanisms that may impact results

prescribe healthy life intervention

moderate–>good performance on adoption, reach and implementation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

outcome eval

A

Measures whether intervention achieved objectives

Mainly quan

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

outcome eval research

A

Ebert et al. (2018)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Ebert et al. (2018)

A

internet and mobile stress management intervention and RCT

intervention v control

int = 7 sessions of problem solving and emotion regulation techniques

baseline v 6 months

cost-effective and lead to cost savings

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

stages of evaluation research

A

Dehar et al. (1993)

Nutbeam (1998)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Dehar et al. (1993)

A

formative - develop and improve programmes at an early stage

process - info on programme implementation, interpretation of outcomes and guiding future research

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Nutbeam (1998)

A

issues with definition and measurement of outcomes and use of eval methods

most powerful interventions = LT

technical problems from scientific rigour and advantages of less well defined content

combine and quan and qual

evals tailored to intervention

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

cause and effect

A

Want to determine whether a cause-effect r’ship exists between intervention and outcome

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

logic of causal inference

A

Under what conditions may we infer that a change in the DV (PA) was really caused by IV (intervention) and not by something else (envs etc.)

What are some of the most plausible rival explanations, and how do we rule them out?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Logic of causal inference research

A

Rothman and Greenland (2004)

Hill (1975)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Rothman and Greenland (2004)

A

causality often debated

more general conceptual model required

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Hill (1975)

A

https://journals.sagepub.com/doi/pdf/10.1177/003591576505800503)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

criteria for inferring causality

A

temporal r’ship

plausibility

strength of association

dose-response r’ship

reversibility

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

temporal r’ship

A

Cause (intervention)must precede effect (increase in PA)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

plausibility

A

Association plausible, and more likely to be causal, if consistent with other knowledge

E.g. reasonable to expect people who receive exercise intervention will increase PA by greater amount than people without intervention

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

strength of association

A

Strong association, as measured by effect size or relative risk, more likely to be causal than weak association

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

dose-response r’ship

A

Occurs when changes in level of cause associated with changes in prevalence/incidence of effect

Greater adherence to exercise intervention (greater dose) the greater change in fitness

Incremental change in outcome, related to dose of intervention

1 intervention - adherence = linked to dose

24
Q

reversibility

A

When removal of cause (intervention) results in return to baseline in outcome (PA), likelihood of association being causal strengthened

25
Q

Internal validity

A

High internal val means that diffs observed between groups related to intervention tested in trial

Means for example, that change in PA in study popn attributed to intervention and not to other factors, such as age, sex/social status

IV has to be only thing influencing DV

Extent to which you’re able to say that no other variables, apart from the IV, caused a change in the DV

Control and treatment group - treat them the same - see whether other variables influencing results

26
Q

Internal validity research

A

Halperin et al. (2015)

27
Q

Halperin et al. (2015)

A

internal val - degree of control exerted over confounds to reduce alternative explanations for results

e.g. controlling ex protocols, prior training, nutritional intake, age, gender etc.

less well controlled: instructions given, verbal encouragement, no. of observers and mental fatigue

28
Q

External validity

A

Described extent to which results of experiment of intervention can be generalised to target/general popn

The extent to which the results of a study are generalisable to other situs/groups

Good sampling needed
- Representative of wider popn

  • Large enough to have adequate power
  • Exclusion criteria should relate to Q of interest
29
Q

External validity research

A

Leviton (2017)

30
Q

Leviton (2017)

A

need greater focus on external val

goal is applicability

methods

  • better description of int
  • combining of stat tools and logic to draw inferences about samples
  • sharper definition of theory
  • more systematic consultation of practitioners
31
Q

Validity research

A

Slack and Draugalis (2001)

Blackman et al. (2013)

32
Q

Slack and Draugalis (2001)

A

establishing internal val of study based on logical process

  • provided by reports structure
  • methods describes how threats minimised
  • discussion assesses influence of bias

threats: history, maturation, testing, instrumentation, regression, selection, exp mortality and interaction

cog map used to guide when addressing threats
- logical description of internal val problems

threats source of extraneous variance

external val addressed by delineating inclusion and exclusion criteria, describing subjects in terms relevant variables and assessing generalizability

33
Q

Blackman et al. (2013)

A

mobile health intervention

see how well inform generalizability

majority RCTs

few studies addressed reach and adoption

research designs need to focus on validity (internal and external)

34
Q

Types of evaluation design

A

Experimental

Quasi-experimental

Non-experimental

35
Q

Experimental

A

Compares intervention with non-intervention

Uses controls that are randomly assigned

36
Q

Experimental research

A

Sanson-Fisher et al. (2007)

Bernal et al. (2018)

37
Q

Sanson-Fisher et al. (2007)

A

RCT often has limitations

alternatives investigated

38
Q

Bernal et al. (2018)

A

control to minimise confounds

consider confounds before deciding on control

39
Q

quasi-experimental research

A

Reeves et al. (2017)

Handley et al. (2011)

40
Q

Reeves et al. (2017)

A

created checklist for quasi-exp designs

  • clustering of int as aspect of allocation/due to intrinsic nature of the delivery of intervention
  • for whom and when outcome data available
  • how intervention effect estimated
  • principle underlying control for confounding
  • how groups formed
  • features of study carried out after designed
  • variables measured before intervention
41
Q

Handley et al. (2011)

A

stepped-wedge and wait-list cross-over design

relevant design features

  • creation of cohort over time that collects control data but allows all Ps to receive intervention
  • staggered intro of clusters
  • multiple data collection points
  • one-way cross over into intervention arm

practical considerations

  • randomisation v stratification
  • training run in phases
  • extended time period for overall study completion
42
Q

non-experimental research

A

Thomson et al. (2001)

43
Q

Thomson et al. (2001)

A

systematic review of exp and non-exp housing intervention studies

able to demo health gains even via non exp interventions

some issues with generalizability

44
Q

experimental examples

A

RCT

Pre-post design with randomised control group is one example of RCT

45
Q

experimental strengths

A

Can infer causality with highest degree of confidence

46
Q

experimental challenges

A

Most resource-intensive
Requires ensuring min extraneous factors
Challenging to generalise to “real world”

47
Q

quasi-experimental

A

Compares intervention with non-intervention

Uses controls or comparison groups that are not randomly assigned

48
Q

quasi-experimental examples

A

Pre-post design with a non-randomised comparison group

49
Q

quasi-experimental strengths

A

Can be used when you are unable to randomise a control group, but you will still be able to compare across groups and/ time points

50
Q

quasi-experimental challenges

A

Diffs between comparison groups may confound
Group selection critical
Moderate confidence in inferring causality

51
Q

non-experimental

A

Don’t use control/comparison groups

52
Q

non-experimental examples

A

Case control (post-intervention only): retrospectively compares data between intervention and non-intervention groups

Pre-post with no control: data from one group compared before and after training intervention

53
Q

non-experimental strengths

A

Simple design, used when baseline data and/ comparison groups not available and for descriptive study

May require least resources to conduct eval

54
Q

non-experimental challenges

A

Minimal ability to infer causality

55
Q

Evaluation in a nutshell (Bauman and Nutbeam, 2014)

A

see notes