Week 5 Random Error Flashcards
What is random error?
= Error that results from data divergence, due to chance alone, of an observation on a sample of the true population
It can never be completely eliminated - but steps can be taken to reduce it.
Careful measurement of exposure and outcome and larger sample sizes
What is the significance of Random Error?
In order to draw meaningful conclusions from study data, we need to be confident that the findings are not due to random error
We use statistical testing to estimate the probability that the results are due to random error (chance)
How can we measure random error?
Hypothesis testing: Statistical tests that reflect the “statistical significance” of a finding. E.g. Creating a null hypothesis which we want to disprove, with a P interval of <0.05
Or through estimation of the range of values that are likely to contain the true value: Confidence intervals (e.g. the 95% convidence interval)
Explain Hypothesis testing
When you perform a test of statistical significance you either reject or accept the null hypothesis. The null hypothesis is that there is not a significant difference between the two groups.
The alternative hypothesis is that there IS a statistical difference between the two groups.
The null hypothesis can only be rejected if there is significant statistical evidence: this is reflected by the P values.
P value regards a statistically significant result as being 0.05 then the null hypothesis cannot be rejected
If P=<0.05 then we can reject the null hypthesis because there is a significant difference
What is Type I error?
Type I error = when the conclusion states there is a significant difference when there isn’t
I.e. the study would incorrectly reject the null hypothesis
Type II error?
When the conclusion states that there is not a significant difference when there actually is one
I.e. the study would incorrectly accept the null hypothesis
What is Alpha a?
The likelihood of producing a Type I error = Alpha(a) error.
It is expressed by the P value.
I.e. the p value is the level of Alpha(a) likelihood that we are willing to accept.
The P value may vary depending on the consequences of making a type I error
What are the considerations for adopting a higher or lower P value?
The P value should be set lower for higher-risk studies.
E.g. if there is no known safe treatment for a severe disease, and you are trialling a treatment - may have more to gain by adopting a less stringent P value.
If there are already safe treatments available for a disease, and you are trialling a more dangerous but potentially more effective drug, then you should adopt a more stringent P value
What is the relationship between Statistical and Clinical Significance
There is not necessarily one.
E.g. there may be found to be a statistically significant difference in performance in cognition tests after taking a certain drug. However, clinically, this may not actually affect patient outcomes. E.g. doing 0.8% better on a test may be statistically proven to be a real difference, but that ‘real’ difference may have no bearing on clinical outcomes
What is the statistial approach to hypothesis tensting?
Test the null hypothesis that there is no significant difference
Null hypothesis is generally rejected if p=. And the P value is the likelihood of Alpha(a) - thus p value = likelihood of type I error
What are the factors that influence P?
1) The magnitude of the main effect
The larger the difference between the groups, the more likely P is to be significant
2) The number of observations
Observations of larger numbers of participants are more likely to reap lower P values (statistically significant) than smaller study sizes
3) The spread in the data: the standard deviation
Wider distribution = higher P value (less significant)
What are the shortcomings of P values?
They do not reflect systematic errors.
Although random error is important, and impossible to completely eliminate, the major concern in most studies is the presence of systematic errors
What is systematic error?
Errors in the kinds of participants involved, measurement-taking biases? Observational errors (e.g. observed associations being confounded by other factors)
What are estimations of true significance in a study results?
Point Estimates
Confidence Intervals
What is a point estimate?
The effect size in the observed study
Effect size = a calculation (scale of 0-1) that can show whether or not a difference is meaningful
I.e. whether a statistical significance is CLINICALLY significant
The ‘effect size’ observed in a study - e.g. the RR, or OR
But it will always be higher or lower than the true effect because of random error