Key principles of epidemiology Flashcards

1
Q

define frequency

A

number of occurrences of an outcome within a given time period

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

two main approaches to epidemiological study

A

descriptive and analytical

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

what information does descriptive epidemiology provide?

A

distribution of health outcomes by age, population type, geography or over time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

what information does analytical epidemiology provide?

A

investigates which factors may be responsible for increasing or decreasing the probability of an outcome

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

difference between sufficient and component cause

A

sufficient cause refers to a factor or sort of factors that produce the outcome. the factors that form a sufficient cause are called component causes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

for an association to be causal, the …. must occur before the ….

A

exposure then outcome

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what can effect causality? (alternative explanation for an association)

A

chance, bias and confounding

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

what is chance and how do you reduce it?

A

possibility that there is a random error and reduced by increasing sample size, using random selection or randomisation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what is bias

A

refers to systematic differences between comparison groups which may misrepresent the association being investigated

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

what is confounding

A

caused when another factor, independently associated with both the outcome and exposure, influences the association being investigated

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

two approaches to analytical epidemiology

A

observation or intervention

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

difference between observational and intervention study

A

observational compares the frequency of an outcome in groups or individuals with and without the exposure of interest. invervention is an experiment - evaluates the effect of reducing a risk factor or increasing a protective factor on the frequency of an outcome.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

define prevalence

A

the number of existing cases in a defined population at a defined point in time divided by the total number of people in that population at the same point in time - it is a proportion

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

how do you measure prevalence?

A

population surveys or cross sectional studies

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

define incidence

A

frequency of new cases in a defined population during a specified time period

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

how do you measure incidence?

A

ecological or cohort studies

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

three ways to consider incidence

A

risk, odds and incidence rate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

define risk

A

total number of new cases in a specified time period divided by the total number of individuals at risk in the population at the start of the time period

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

how do you work out secondary attack rate

A

number of new cases among contacts in a specified time period divided by total number of contacts of a primary case in that time period

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

define odds

A

ratio of two proportions so number of new cases in a specified time period divided by number who did not become a case during that time period

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

what do you use if it is an open population rather than a closed population to work out incidence rate?

A

population time at risk

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

examples of relative measurers that use ratios

A

prevalence ratio, risk ratio, odds ratio, incidence rate ratio

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

what does relative risk measure

A

how much more likely it is that an exposed individual will develop the outcome compared with the unexposed individual

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

what studies do you use to calculate risk ratio?

A

ecological, cohort or intervention

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
how do you calculate odds ratio?
odds of outcome in exposed group divided by odds of outcome in unexposed group
26
what studies do you use to obtain incidence rate ratio?
ecological, cohort or intervention
27
when would you use incidence rate ratio over the odds ratio of exposure?
where people are entering and leaving the study population or have changing levels of exposure
28
if the outcome is rare, are the measures of association more similar or different?
more similar
29
when would you use attributable risk?
when you want to know the excess incidence of the outcome that we an attribute to the exposure (if we assume a causal relationship)
30
how do you work out attributable fraction?
attributable risk divided by incidence in exposed OR relative risk - 1 divided by relative risk
31
what is the opposite of attributable fraction? (i.e. if there's a protective factor)
preventable fraction
32
how to work out the population attributable fraction?
the incidence in the population minus the incidence in unexposed divided by the incidence in population
33
define p value
the probability that an observed value or association from a sample occurred by chance alone and that it does not exist in the population from which the sample was selected
34
define confidence interval
the range of values, estimated from a sample, within which the true population value is likely to be found
35
if confidence interval includes 1 what can we say about the association?
it is not significant
36
how do you reduce bias?
through random selection of study participants or random allocation of individuals to comparison groups
37
different types of bias
selection | information
38
when does selection bias occur
when there is a systematic difference between the characteristics of individuals samples and the population from which the sample is taken or a systematic difference between comparison groups
39
what factors affect whether selection bias occurs?
the definition of the study population or comparison group the inclusion and exclusion criteria the rate of loss to follow up healthy worker
40
when does information bias occur
when there is a systematic difference between comparison groups in the way that data is collected
41
how can information bias be introduced?
by those measuring the outcome (observer bias), the study participants (responder bias) or measuring tools (measurement bias) or misclassification
42
when does non differential misclassification occur?
occurs when both comparison groups are equally likely to be misclassified - unable to measure effectively. causes comparison groups to appear more similarly than they actually are and may lead to underestimation of the strength of association. independent of exposure or outcome
43
when does differential misclassification occur?
when classification of the exposure is dependent on the outcome or vice versa - due to observer or respondent bias
44
how do you reduce the effect of observer bias?
blinding, objective vs subjective measures
45
three ways to avoid confounding
randomisation - of individuals to exposure and control groups restriction - limits the study to people who are similar in relation to the confounder matching - selects two comparison groups to have the same distribution of potential confounders
46
how to control for confounding in the analysis
stratification - extension of frequency matching as it measures the association between exposure and outcome separately for each category of confounder statistical modelling - allows us to adjust simultaneously for several confounders using multivariable regression analyses.
47
bradford hill criteria for building up evidence for a causal relationship
strength - the stronger the association between the exposure and outcome, the less likely the relationship is due to another factor consistency - repeatability of the result temporality - exposure comes before the outcome dose-response - increased risk of outcome with increased exposure plausibility - existence of a reasonable biological mechanism reversibility - whether an intervention to remove the exposure results in the elimination of the outcome coherence - consistency with other information analogy - similarity between other established cause-effect relationships helps support the argument specificity - the relationship is specific to the outcome of interest
48
ecological studies: what do they do
compare frequency of outcome an exposure at a population level
49
cross sectional studies: what do they do
compare prevalence of outcome with exposure status at one time point from a random sample of individuals
50
cohort studies: what do they do
compare the incidence of outcome in individuals with recorded differences in exposure
51
case control: what do they do
select individuals on the basis of their outcome status and analyse whether they differ in relation to previous exposure
52
intervention studies: what do they do
allocate a protective exposure and compare outcome between those exposed and unexposed
53
Ecological study: what measures do you use? advantages and disadvantages?
prevalent or incident cases easy to collect data, rapid, inexpensive cannot show causality, high probability of confounding, ecological fallacy
54
cross sectional studies: what measures do you use? advantages and disadvantages?
prevalent cases, prevalence ratio easy to collect data, rapid difficult to know if exposure preceded outcome, selection bias, confounding
55
cohort studies: what measures do you use? advantages and disadvantages?
incident cases - risk ratio, odds ratio or rate ratio low probability of selection bias, recall bias and confounding loss to follow up, more expensive, more difficult as large number of participants required
56
case control: what measures do you use? advantages and disadvantages?
no measures of outcome frequency, odds ratio of exposure low probability of loss to follow up high probability of selection and information bias, confounding
57
intervention: what measures do you use? advantages and disadvantages?
prevalent or incident cases, preventable fraction low probability of selection and information bias and confounding, satisfy temporality consideration, satisfies reversibility, can radnomize subjects to reduce effects of chance, reduce information bias loss to follow up, more complex, expensive
58
difference between statistical power and precision
statistical power = probability of detecting an effect if it is real statistical precision = probability of detecting an effect if it is not real
59
different types of random sampling
simple random sampling systematic sampling - regular intervals stratified sampling - equal proportion from each group multi stage sampling - sampling frame required for high level unit and for primary sampling units within higher levels sampled cluster sampling - also uses organisational structure but samples all or most of the primary sampling units
60
if you need to exclude certain individuals what do you use when sampling?
exclusion and inclusion criteria
61
what is informed consent?
participants need to give their consent to participate
62
medical records, census data, health surveys, schools records are examples of what type of data collection method?
indirect data collection methods
63
questionnaires, structured interviews and clinical examination are examples of what type of data collection method?
direct data collection methods
64
why do we use unique identifiers?
so we can link all necessary data without identifying participant
65
three types of quantitative variables
binomial - two values categorical continuous
66
how do you usually present a meta analysis
forest plot
67
what does a meta analysis do?
weights the individual estimates in relation to the sample size of each study
68
when do we use direct standardisation?
in ecological studies to reduce confounding when we know the group specific outcome rates in the study population. the age specific rate is the incidence rates calculated separately for each group - can then add them up to give a standardised rate
69
when do we use indirect standardisation?
calculates the expected number of cases in the study population if the age specific rates of a standard population had applied. used if we don't know the age specific rates. the expected number of cases can then be used to calculate a standardised mortality ratio which compares the observed number of deaths to that expected in the same population.
70
what is ecological fallacy?
in ecological studies when a mismatch arises from trying to draw conclusions about individual level epidemiological associations from a group level study
71
what is mixing in ecological study interpretation?
geographical comparisons may suffer from migration of populations between groups
72
if 95% confidence interval does not include 1, what can you say about the p value?
p value will be less than 5 (less than 5% chance that this association is not true)
73
difference between point and period prevalence
point prevalence is at a specific point in time and period prevalence is over a period of time. period prevalence increases possibility of recall bias
74
what are cross sectional studies at high risk of when assessing causation?
reverse causality
75
are analytical cohort studies retrospective or prospective?
can be either as long as cohort defined on exposure and not outcome
76
in a cohort study, if follow up times are similar what frequency measure do we use? if follow up times are different what do we use?
risk if similar and rate if follow up times differ
77
what is the issue with case control studies and bias?
susceptible to information and selection bias
78
how can you reduce confounding in case control studies?
identify controls with the same characteristics as the cases - matching
79
what are nested case control studies?
cases are members of a cohort that have developed the outcome and control are those without the outcome. can automatically match on factors that are common to all cohort members
80
in a case control study what do you need to ensure when sampling the population?
that they are representative of the target population in relation to the frequency of exposure, selection of cases and control are not influenced by their exposure
81
why might inclusion of prevalent cases in a case control study be an issue?
inclusion of prevalent cases, especially chronic outcomes, can create problems for determining exposure to certain risk factors that may change over time which may lead to reverse causality. may also lead to underrepresentation of severe cases
82
why can't case control studies directly estimate prevalence or incidence of the outcome or frequency of the exposure in the general population?
because they don't randomly sample the population, instead they select individuals based on their outcome status
83
what is the ideal intervention study?
randomised controlled trial
84
difference between efficacy and effectiveness?
``` efficacy = measures the effect of the intervention under "experimental" conditions, where maximum effort is put into intervention delivery. once proven to be efficacious... effectiveness = measures the effect of the intervention of the intervention under routine conditions. ```
85
what is a plausibility study?
evaluates the effectiveness of interventions by comparison with a historical, geographical or opportunistic control group but without randomisation.
86
when might you use a plausibility study?
when an intervention is so complex that RCT results will be unacceptably artificial. when an intervention is known to be efficacious or effective in small scale studies but effectiveness on a large scale must be demonstrated when ethical concerns prevent the use of an RCT
87
what is a plausibility study with a historical control group?
not possible to have a contemporary comparison arm so it may compare outcome frequency before and after the intervention is introduced. makes it difficult to distinguish between the effect of the intervention and any other changes that may have affected the outcome over the study period.
88
what is a plausibility study with a geographical control group?
the use of communities or individuals as contemporary control outside the trial area may adjust for the effects of temporal changes during the course of the evaluation. however, there are likely to be differences between intervention and control. can use stepped wedge design (by introducing into population in phases)
89
what is a plausibility study with a opportunistic control group?
involves the use of individuals or communities that should have reached the intervention but did not because the programme was unable to reach them.
90
define RCT
the existence of a contemporary comparison group of study subjects who do not receive the intervention known as the "control" and the random allocation of study subjects to the intervention and control arms
91
how is a cluster randomised controlled trial different?
groups of individuals are randomly allocated to the intervention and control arms
92
what happens in a factorial trial?
two or more interventions are compared individually and in combination against a control comparison group
93
what happens in a crossover design RCT?
each trial subject acts as its own control by receiving either the intervention or control at different points in the study, with a washout period to avoid contamination between the study periods
94
different types of random allocation
``` simple randomisation systemic randomisation blocked randomisation stratified randomisation matched pair randomisation ```
95
what is simple randomisation?
uses number tables or a computer generated random number list
96
what is systematic randomisation?
allocates participation to each group alternately but subject to selection bias
97
what is blocked randomisation?
restricts the allocation list to ensure equal numbers in all study arms.
98
what is stratified randomisation?
divides participants into subgroups or strata based on key risk factors and equal numbers of subjects from each stratum are randomly allocated to each study arm. this ensures that suspected confounders or effect modifiers of the association between intervention and outcome are equally distributed between comparison groups
99
what is matched pair randomisation?
form of stratified random allocation that matches individuals or communities into pairs with similar baseline risks of the outcome
100
what is the primary analysis of an intervention study?
intention to treat
101
what is the efficacy of an intervention?
the proportion of cases that can be prevented by the intervention - also known as protective efficacy or preventable fraction and calculated by doing 1 - relative risk
102
what is primary prevention and example?
aims to stop an outcome developing by preventing or reducing exposure to a risk factor vaccination against an infectious disease
103
what is secondary prevention and example?
aims to interrupt progression from early to mid-stage of the outcome by early detection and prompt treatment. e.g. blood lead screening in USA
104
what is tertiary prevention and example?
aims to reduce complications or severity once outcome is established and symptomatic by offering appropriate treatment and intervention. e.g. by monitoring blood glucose of those with diabetes to prevent renal disease and glaucoma
105
what is screening used for?
to identify small numbers of individuals at high risk of an outcome
106
how do you work out sensitivity?
the proportion of people who have the disease who are correctly identified.
107
how do you work out specificity?
the proportion of people who do not have the condition that are correctly identified.
108
what is the positive predictive value?
the likelihood of an outcome based on the result
109
what is the negative predictive value?
the likelihood of no outcome based on a result
110
disadvantages of screening
``` false positives and false negatives potential harm (X-ray in mammography) good quality control needed specialist equipment may be required staff may require additional training ```
111
WHO criteria for screening
the condition should be an important health problem natural history of the condition should be well understood there should be a recognisable latent or early stage there should be a suitable method for detection screening method should be acceptable to population should be an accepted treatment for those with the condition should be an agreed policy on who to treat facilities for diagnosis and treatment should be available costs of detection should be balanced inc elation to overall healthcare spending screening should be ongoing and not one off
112
what is lead time bias?
screening identifies an outcome earlier than it otherwise would have been identified but has no effect on outcome (so looks like prolonging)
113
what is length time bias?
outcomes that take longer to develop to a stage where they threaten health - more likely to pick up less aggressive tumours with better prognosis leading to over-estimate of screening success.
114
what is the prevention paradox?
where the majority of cases of a disease come from a population at low or moderate risk of that disease, and only a minority of cases come from the high risk population (of the same disease - target whole population
115
what is a case definition?
set of standard criteria for determining whether or not a person has a particular disease or health related event
116
what effect does an emerging new treatment that increases the life expectancy of HIV cases have on prevalence/incidence?
increased prevalence
117
what effect does the discovery of a cure for DM have on prevalence/incidence?
decreased prevalence
118
what effect does the introduction of a new malaria vaccine have on prevalence/incidence?
decreased incidence
119
different types of bias
``` information selection recall sampling non response ```
120
difference between descriptive and analytical cross sectional studies
descriptive - collect info on frequency and distribution of health related exposures or outcomes in a defined population analytical - investigate the association between exposure to risk factors and outcome of interest
121
examples of when selection bias can occur in cross sectional, cohort, case control and intervention studies
cross sectional: non response cohort: healthy worker effect, loss to follow up case control: selection of cases and controls intervention: systematic selection for intervention/controls
122
what is validity?
the degree to which a measurement measures what it purports to measure
123
what is reliability?
the degree to which the results obtained by a measurement procedure can be replicated