Session 2a Flashcards
Hypothesis
Empirically testable statement derived from our theories.Hypotheses are assumptions made about population parameters (not sample statistics).
Hypothesis testing
Traditionally, experimental research engages in a procedure for hypothesis testing: Null hypothesis significance testing (NHST). NHST is still widely used. More recent approaches focus on effect sizes and formation of confidence intervals
Hypothesis testing step 1
- Step 1: Set Up a hypothesis (Usually a prediction that there is an effect of certain variable(s) in the population)
- Step 2: Choose α (significance level: decide the area consisting of extreme scores which are unlikely to occur if the null hypothesis is true)
- Step 3: examine your data and compute the appropriate test statistic
- Step 4: make the decision whether to reject or not reject the null hypothesis
Step 1 (hypothesis)
- Null Hypothesis (H0): This is what we test statistically. Null is true: no effect.
- Alternative Hypothesis (H1): research/experimental hypothesis. Non-directional H1: Some effect. Directional H1: Specifies the direction of the difference.
Step 2 (criterion)
α is the proportion of times we are willing to accidentally reject H0, even if H0 is true. The cutoff value of the test statistic for α is called the critical value.
Step 4 (decision)
Compare the calculated value of your test statistic to the critical value for α. If your value is greater than or equal to the critical value, reject H0. Otherwise, retain H0. A decision to reject H0 implies support for H1.
Step 4 alternative (p-value)
Alternatively, look at the significance level (p-value) for the test statistic value. The p-value represents the proportion of data sets that would yield a result as extreme or more extreme than the observed result if H0 is true. Such values often given by SPSS, R, or other statistical software.. If p ≤ α (e.g., p ≤ .05), reject H0. Otherwise, retain H0. It is a Yes-No decision.
Significance
If H0 is rejected, you may conclude that there is a statistically significant effect in the population.
A “statistically significant” effect does not indicate that. . .
- We have a precise estimate of the effect: it may be that an effect is “significant”, but there is some error around our estimate. The effect in the population may be smaller or larger than our estimate. The amount of error is represented in the standard error for the estimate.
- The effect is important or meaningful:
Suppose we find that seeing a spider increases anxiety by 2 points. Is 2 points a meaningful increase? The anxiety gain may be “significant” if it was observed from many people, but it does mean that the effect is practically meaningful.
Confidence interval
A Confidence Interval gives us information about the precision of our estimates.
Example: a 95% confidence interval (CI) may indicate that true increase in anxiety in the population is between 0.5 and 3.5 points. If we repeated our experiment many times, 95% of the time a 95% CI will contain the true effect. As sample size increases, our estimate becomes more precise (and our CI intervals become more narrow). As α decreases, our CI intervals become larger or wider.
Effect sizes
A standardized measure of the magnitude (importance) of a treatment effect. Commonly used measures of effect size:
- Pearson’s correlation coefficient or correlation ratio squared
- Cohen’s d
- Omega or omega squared
- Eta (η) squared
It is strongly encourage to always report effect size measures.
Errors in hypothesis testing
The procedures we follow in hypothesis testing does not guarantee that our decision will be correct. Two types of errors:
- Type I: Reject H0 when it is true (False Positive). α = probability of committing Type I error
- Type II: Retain H0 when it is false (False Negative). β = probability of committing Type II error
- 1 − β (power) = probability of correctly rejecting a false H0
Controlling errors in hypothesis testing
Complete control of α and β is not possible because of the relationship between α and β. Higher values of α means lower values of β (i.e., a less conservative test), which means the power to reject H0 is higher.
We usually select α a priori. The most common choice is α = 0.05, which means that if H0 is true, we stand only a 5% chance of rejecting it (and thus making a Type I
error).
z-test for a single mean
- Purpose: Based on the sample mean (X) we test whether the population mean (µ) is equal to some hypothesized value
- Prior Requirements/Assumptions: The variable, X, in the population is normally distributed. The sample must be a simple random sample of the population (independence of observations). The population standard deviation, σ, must be known.
- Computing a z-test for a single mean is like calculating a Z-score for your sample mean.
Limitations of z-test
- Knowing the true value of the population standard deviation (σ) is unrealistic, except in cases in which the entire population is known
- Alternative? The t-test