2. Confidence Intervals and Logic Of Hypothesis Testing Flashcards
What makes a statement scientific?
It has to be testable and falsifiable
If we’d collected 100 samples, calculated the mean and then calculated a confidence interval for that mean, then for 95 of these samples the confidence intervals we constructed would…
contain the true value of the mean in the population
What does a significant test statistic tell us?
That the test statistic is larger than we would expect if there were no effect in the population.
Of what is p the probability?
p is the probability of observing a test statistic at least as big as the one we have if there were no effect in the population (i.e., the null hypothesis were true).
A Type I error occurs when
We conclude that there is an effect in the population when in fact there is not
(False positive)
A Type II error occurs when
We conclude that there is not an effect in the population when in fact there is.
(False negative)
Power is the ability of a test to…
detect an effect given that an effect of a certain size exists in a population.
We can use ____ to determine how large a sample is required to detect an effect of a certain size.
power
True or False:
Power is linked to the probability of making a Type II error.
True
What is the relationship between sample size and the standard error of the mean?
The standard error decreases as the sample size increases.
- The standard error is the standard deviation of the distribution of sample means
- the sample mean is closer to the population mean
In general, as the sample size (N) increases, the confidence interval…
gets narrower
In large samples, can small effects be deemed ‘significant’?
Yes
The assumption of homogeneity of variance is met when:
The variances in different groups are approximately equal
- To make sure our estimates of the parameters that define our model and significance tests are accurate we have to assume homoscedasticity (also known as homogeneity of variance)
If the p value is very small then we conclude that the model….
fits the data well (explains a lot of the variance) and we gain confidence in the alternative hypothesis H1
What does effect size tell us?
An effect size is a standardized measure of the size of an effect:
- Standardized = comparable across studies
- Not (as) reliant on the sample size
- Allows people to objectively evaluate the size of the observed effect
Effect size (Cohen’s d):
Small =
Medium =
Big =
Small = 0.20 Medium = 0.50 Big = 0.80
Effect size (Pearson’s r):
Small =
Medium =
Big =
Small = 0.10 Medium = 0.30 Big = 0.50
What does statistical power tell us?
statistical power is the likelihood that a study will detect an effect when there is an effect there to be detected. If statistical power is high, the probability of making a Type II error, or concluding there is no effect when, in fact, there is one, goes down
- It is generally accepted that power should be .8 or greater
Bigger sample size with smaller SD = bigger power
Bigger difference or standard error between H and H1 = bigger power
What is Levene’s test?
Tests if variances in different groups are the same
Significant = Variances not equal
Non-Significant = Variances are equal
Ways of reducing bias in the data:
Trim the data:
- Delete a certain amount of scores from the extremes
Windsorizing:
- Substitute outliers with the highest value that isn’t an outlier
Analyze with Robust Methods:
- Bootstrapping
Transform the data:
- By applying a mathematical function to scores.
(log transform, square root transform & reciprocal transform)