Critical appraisal course Flashcards
What is EBM
The conscientious, explicit and judicious use of current best evidence in making decisions about the care of a patient
5 steps of EBM
- Clinical question
- EVidence
- Critical appraisal
- Application
- Implementation and monitoring
Stages of critical appraisal
- The clinical question
- Methodology: study design, recruitment, variables and outcomes
- Results: data analysed and differences between groups assessed for significance
- Applicability
Internal and external validity
Internal: extent to which the results from the study reflect the true results
External: extent to which study results can be generalised
Efficacy and effectiveness
Efficacy is the impact of interventions under optimal (research) setting
Effectiveness is whether the interventions have the intended or expected effect under ordinary clinical settings
The efficacy of an intervention is almost always better than the effectiveness
True
The acceptance of EBM means all clinicians practice the same
False
Patient values have no role to play in EBM
False
The effectiveness is almost always better than the efficacy
False
A research project taking place in the outpatient clinic is almost always going to give effectiveness data rather than efficacy data
True
An RCT can answer any type of clinical question
False
The clinical question determines which study designs are suitable
True
A clinical question can usually only be answered by one type of study design
False
The critical appraisal should start by examining the study design
False
Broad categories of studies and what they can achieve
- Observational descriptive
- survey, qualitative, case report/series
- Generate hypothesis - Observational analytical
- case-control (outcome->exposure)
- cohort (exposure->outcome)
- test hypothesis - Experimental
- RCT, crossover, N of 1
- intervention - Others
- Ecological: information about the population
- Pragmatic: real life environment. Example- all people in clinical location, outpatient, randomised to receive particular treatment. More reflective of everyday practice
- Economic
- Systematic review/MA
Case series is observational analytic
False
Case control is observational descriptive
False
Qualitative study
Opinions are elicited from a group of people with emphasis on subjective meaning and experience. Complex issues can be identified
Data gathering and data analysis develops iteratively-> results inform further samplig
Inductive-> knowledge generated through data sampling and gathering as opposed to other scientific methods where research is deductive
Other names for case control
Retrospective or case comparison
Case control- type, advantages and disadvantages
People with outcome variable, are compared to those without outcome variable, to determine risk factors they have been exposed to in the past
Adv: cheap and easy good for rare outcomes few subjects required good for diseases with long duration between exposure and outcome
Dis:
not for rare exposures
Recall problems
Control groups can be difficult to select
Cohort study- design, retrospective type, other names, adv, disadv
A group of people with exposure are followed up to see the development of an outcome
Also called prospective or follow-up
Retrospective type is using cohort data that already exists (say from 20 years ago, with exposure)
Adv: Good for rare exposures multiple outcomes temporal relationship estimation of outcome incidence rates
Dis: May take ++time from exposure to outcome expensive attrition rates unsuitable for rare outcomes
Recall bias is a bigger issue fir CC or Cohort
Case control
Which is better study design for rare exposures
Cohort
Whch study design is betten when long time from exposure to outcome
CC
Disadvantages of RCT
Expensive
Time consuming
When might you use a crossover trial
When unable to get enough subjects for RCT
Given one intervention, then switched half way through
Disadvantages of cross-over trials
The order f interventions might be important Carryover effects (drugs with long half lives) or prolonged discontinuation/withdrawal May be difficulty using historical controls if conditions were different
N of 1 trials
Experimental version of case report
Single person
Given randomised treatments
Report on response
Historical control bias is an issue in which type of study design
Crossover
Which types of biases are an issue in open label
Selection and observation bias
Two sources of methodological error
Bias
Confounding factors
Definition of bias
Any process at any stage of inference, which tends to produce results or conclusions that differ (systematically) from the truth
Not by chance
Researchers must try to reduce bias
categories of bias
- Selection bias= recruitment of sample
- Performance bias= running of the trial
- Observation bias= data collection
- Attrition bias
2+3= measurement bias
Bias versus confounding
Confounders are real life relationships between variables that already exist, and so are not introduced by the researcher
Selection bias
Error in recruitment of sample population
Introduced by: Researchers= sampling bias - admission or berkson - diagnostic purity - neyman bias - membership bias - historical control Subjects= response bias - volunteers differ in some way from the population - example if more motivated to improve their health, and adhere ++to the trial
Types of sampling bias
- Berkson bias
- sample taken from hospital setting, and therefore rates/severity of condition is different compared to target population - Diagnostic purity bias
- co-morbidity are excluded from sample population, and therefore does not reflect the complexity in the target population - Neyman bias
- prevalence of condition, does not reflect the incidence, due to time gap between exposure and actual selection. such that some with exposure are not selected (have died)
- example, giving treatment following MI. Some may die shortly after MI and therefore not be selected, therefore there is better prognosis already - Membership bias
- members of group selected may not be representative of target population - Historical control bias
- subjects and controls chosen across time, so definitions, exposures, diseases and treatments may mean they cannot be compared to one another
Protecting against performance bias
Systematic differences in the care provided, apart from intervention evaluated
Standardisation of care protocol and blinding protects
Randomisation and blinding
Types of observation bias
Failure to measure or classify the exposure or disease correctly. Can be due to researcher or participant.
Researcher
- Interviewer (ascertainment) bias
- when researcher not blinded, may approach subject differently depending on if they know they are taking the treatment or the placebo - Diagnostic/exposure suspicion bias
- Implicit review bias
- Outcome measurement bias
- Halo effect- knowledge of patient characteristics influences the impression of patients with respect to other aspects
Subject
- Recall bias
- Response bias- answers questions in a way they think the researcher would want
- Hawthorne effect- behaving in a way, usually positively, as aware being studied
- Social desirability bias
- Bias to middle and extremes
- Treatment unmasking
Attrition bias
The numbers of individuals dropping out differs significantly between the groups
Those left may not reflect the sample or target population
Intention to treat analysis will need to be conducted
Bias occurs by chance
F
lack of blinding could lead to ascertainment bias
T
What is a confounder, positive and negative
When there is a relationship between two variables, that is attributable to, or confounded by the presence of a third. May make it seem like the two variables are associated when they are not (positive confounder eg coffee and lung cancer *smoking, overestimating association), or fail to show association when there is (negative confounder poor diet and CVD *exercise, underestimating association)
To be a confounder
- Must be associated with the exposure, but not the consequence
- The outcome, independently from the exposure
Controlling for confounders
- Restriction
- inclusion and exclusion criteria - Matching
- Randomisation
Accounting for confounders using statistical methods
- Stratified analyses
- can only control for a few - Multivariable analysis
- Accounts for many, need at least 10 subjects, in a logistic regression
If matching was done, McNemar test or conditional logistic regression
Simple randomisation
Subjects are randomised to groups as they enter the trial
Selected independently of each other
Block randomisation
Differs from simple randomisation in that subjects are not allocated independently
Subjects are assigned to “blocks”, which as they fill, then distributed evenly to intervention or control group
Stratified randomisation
Subgroups are formed in relation to a confounding factor, then in each stratum, block randomisation occurs, so the confounders are equally distributed
Concealed allocation and methods
When the treatments being administered in the different arms of the study remain secret
Part of the randomisation process
Methods:
- Centrally controlled
- Pharmacy concealed
- Sequentially numbered, opaque, sealed envelops
- Numbered/coded bottles or containers
Allocating an equal number of subjects in each groups is possible with simple randomisation
Yes- possible by chance
Pygmalion effect (Rosenthal effect)
Subjects perform better than others because they are expected to
Power of positive expectations
George Bernard Shaw- Pygmalion
Placebo effect
In healthcare, the pygmalion effect is often called the placebo effect
Latin meaning of placebo
“I shall please”
2 methods to make expectations equal
- Allocation concealment
- at time of selection - Blinding
- once the subjects start treatment
Problem with single blinding
If the subject is blind, the researcher is still in full possession of the facts
The subject may still be influenced by the behaviour of the researcher who may have expectation about the outcome
Blind assessment
Assessment of the outcome measures during and at the end of the study is made without any knowledge of what the treatment groups are
Placebo factors
Multiple pills
Large pills
Capsules
Reliability definition and subtypes
Consistency of results on repeat measurements by one or more raters over time
- Inter-rater
- level of agreement by 2+ assessors at the same time - Intra-rater
- one rater, same material, different time - Test-retest
- level of agreement from initial test results to repeat measures at later date - Alternate form reliability
- reliability of similar forms of the test - Split-half reliability
- reliability of test divided in two, with each half being used to assess the same material under similar circumstances
Quantifying reliability
Compare the proportion of scores which agree, with the proportion that would be expected to agree by chance
Reliability co-efficient
Kappa (Cohen’s) statistic k
Kappa or Cohens is used to test measures of categorical variables
Inter rater reliability in qualitative
Also known as chance-corrected proportional agreement statistic
Measures the proportion of agreement over and above that expected by chance
If agreement is no more than expected by chance k=0
To be significant, k> 0.7 is normally necessary
Strength of agreement or association
0= chance agreement only <0.2 poor agreement beyond chance 0.21-0.4 Fair agreement beyond chance 0.41-0.6 Mod 0.61-0.8 Good agreement 0.81-1.0 Very good 1 Perfect agreement
Crohnbach’s alpha is
Used in complicated tests with several parts measuring several variables
When you have multiple Likert questions
Internal consistency/reliability
No formal test statistic
>0.5 mod
>0.8 excellent
Intraclass coefficienct
Used for tests measuring quantitative variables, such as BP
Validity and subtypes
Extent to which a test measures what it is supposed to measure
- Criterion- predictive, concurrent, convergent, discriminant
- Face
3, Content - Construct
- Incremental
Criterion validity
demonstrates the accuracy of a measure or procedure by comparing it with another measure or procedure that has been demonstrated to be valid
- Predictive: extent to which the test can predict what it theoretically should be able to predict
- Concurrent validity: extent to which the test can distinguish between two groups it theoretically should be able to distinguish
- Convergent validity: the extent to which the test is similar to other tests that it theoretically should be similar to
- Discriminant validity: the extent to which the test is not similar to other tests that it theoretically should not be similar to.
Other types of validity
- Face validity: superficially looks to measure what it should
- Content: measures variables that are related to that which should be measured
- Construct validity: extent to which a test measures a theoretical concept by a specific measuring device or procedure
- Incremental validity: the extent to which the test provides a significant improvement in addition to the use of another approach. A test has incremental validity if it helps to the use of another approach.
Intention to treat analysis
All the subjects are included in the analyses as part of the groups to which they are randomised, regardless of whether they completed the study or not.
Last observation carried forward, disadvantages
Way of accounting for subjects that drop out before the end
Disadvantages:
1. Underestimation of treatment effects
- intervention expected to lead to an improved outcome
2. Over estimation of treatment effects
Intervention expected to slow down a progressively worsening condition
per protocol analysis
only those subject remaining in the study are used in the analyses
introduces bias through exclusion of participants who dropped out
when is incidence preferred over prevalence and vice versa
When disease is frequent and short duration-> incidence
When long duration, slow, rare-> prevalence more useful to indicate impact of disease on the population
Mortality rate
Type of incidence rate that expresses the risk of death in a population over a period of time
Standardised mortality rate
Adjusted for confounding factors
Standardised mortality ratio
Ratio of observed mortality rate compared to expected mortality rate
Types of data summary
1. Categorical Nominal Ordinal 2. Quantitative Discrete Continuous
Categorical data
No numerical value
Not measured on scale
No in between values
- Nominal= unordered
- binary, dichotomous= only 2 mutually exclusive categories (dead/alive, male/female)
- multi-category= mutually exclusive categories, bearing no relationship to each other (married, engaged, single, divorced) - Ordinal= numbered
- order inherent, but not quantified
- can assume non-parametric
Quantitative data (numerical)
- Discrete
- counts (number of children, asthma attacks) - Continuous
- can have a value within the range of all possible values
(age, body weight, ht, temp)
Another term for normal distribution
Gaussian distributioni
Statistical tests to describe samples with different data samples: categorical, quantitative
Categorical= mode, frequency
Quantitative=
1. Non-normal distributed-> median, range
2. Normally distributed-> mean SD