Lecture 9 Flashcards

Experimental Design

1
Q

Why do we need good experiments?

A

Bad ones waste money, time, resources and effort. They can also have real-world consequences

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Two classes of research design:

A

-Observational/ correlational/descriptive

-Manipulative/experimental

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

O/C/D Research

A

Looks for patterns and/or associations. Correlational; cannot prove causation. e.g. association between an inherited genotype and diabetes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

M/E Research

A

Manipulates conditions. Stronger design; tests cause and effect.Seeks to prove causation. e.g. create transgenic animal model with mutation in the suspect gene, compare glucose metabolism against normal animal

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Experimental design essentials

A

Clear and precise question, Controls/comparator group,
Randomisation and Repetition

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Observational studies

A

Easier to collect data in ‘natural’ settings without interfering with what you’re observing. Easier than manipulative studies. Very often categorical data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Reverse Causation

A

group B has an effect on group A

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

(Third Factor) Confounding variable

A

Group C has an impact on groups A and B

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Why do we need controls?

A

Can’t tell difference between a real treatment effect and a random change over time. Also: Placebo effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Why do we replicate experiments?

A

Individuals may vary

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

RCT

A

The ‘Gold Standard’ study design for evaluating interventions. Randomised (sometimes double-blinded/ placebo) Controlled Trial

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Measurement Validity

A

The degree to which any measurement approach or instrument succeeds in describing or quantifying what it is designed to measure

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is Bias?

A

A systematic error (caused by the investigator or the subjects) that causes an incorrect (over- or under-) estimate of an association

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Most common type of bias

A

Selection Bias. (Control selection bias, Loss to follow-up bias, Self-selection bias, “Healthy worker” effect and Differential referral or diagnosis of subjects)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Double-Blind Study

A

The treatment vs control status is unknown to the researchers as well as to the subjects. Avoids the possibility of both measurement bias and the placebo effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

False Positive

A

Type I Error: We reject the null hypothesis when it is true and in reality there is no difference or association

17
Q

False Negative

A

Type II Error: We accept the null hypothesis when it is false and in reality there is a difference or association

18
Q

Errors with Multiple Comparisons

A

By using a p-value of 0.05 as the criterion for significance (α) we’re accepting a 5% chance of a false positive (of calling a difference significant when it really isn’t)

19
Q

Power and sample size

A

Statistical power is the likelihood of detecting a significant difference or significant association, by sampling, if it actually exists between the underlying populations.
Too few subjects = total waste
Too many subjects = partial waste

20
Q

If sample size is too small…

A

Confidence intervals are too broad, so we are unlikely to detect, with statistical significance, a true difference between means