Replication + peer review + multiple comparisons Flashcards
Why replication could be important
- Publication bias
○ Journals only publish significant results?
○ This is problematic, but unlikely to be the whole story
○ 20 experiments for 1 payoff seems very inefficient (1/20 times you’ll incorrectly reject the null) - Multiple comparisons
○ Type I error rate – the probability of rejecting H0 when true.
○ Decision-wise error rate is 5% – Experiment-wise error rate depends on the number of tests
○ Two hypothesis tests
§ P(reject at least one H0|both H0 true)
Researcher degrees of freedom (hidden multiple comparisons)
- Refer to the flexibility researchers have in designing, conducting, and analysing studies, which can lead to a multiple comparisons problem, even when researchers only perform a single analysis
○ Might trim the outliers
○ Increased false positive rates: Researchers might find statistically significant results even when there is no real effect in the population.
○ Overoptimistic results: The findings might appear more impactful or definitive than they should be, given the potential for bias introduced by the analytical choices.
allows researchers to to present exploratory analyses as confirmatory
Why are confidence intervals better than p-values for guiding replication in a single study?
Confidence intervals show the range of plausible effect sizes and their precision, making it easier to assess consistency in replication.
Confirmatory experiment
- Analysis is planned before data is collected
- Only tests that are directly related to a prespecified test are performed
- Can pre-register these analyses, to protect against researcher degrees of freedom
- Type I error rate is 5%
Exploratory experiment
- Unplanned analyses performed on data set
- Usually many tests are performed
- Researcher degrees of freedom used
- The Type I error rate is not maintained at 5%
What is the journal review process
- Action editors select reviewers based on expertise.
- Action editor reads paper and reviewer comments to decide: accept, accept with minor revision, reject but invite major revision, or reject.
- Reviewers and editors are unpaid; publishers hold copyright and sell journals
Determining journal quality:
- Quality of editors and reviewers.
- Acceptance rate.
- Impact factor (citation rate).
- Reputation.
Australian Research Council classification
What are multiple comparisons in statistical designs?
Multiple comparisons occur when more than one comparison between means is tested in a study, increasing the chance of error
What risk is associated with multiple comparisons?
They can inflate the probability of Type 1 errors (false positives).
How does the experiment-wise error rate relate to the decision-wise error rate?
If the decision-wise error rate is set at α, the experiment-wise error rate will be greater than α when there are multiple comparisons.
Why is controlling the experiment-wise error rate considered more conservative?
It reduces the overall chance of making any Type 1 error across all tests, offering stronger protection against false positives
When is it appropriate to control the decision-wise error rate?
Only when all comparisons are independent of each other
What makes statistical tests independent?
Tests are independent if they do not share means
Are tests of main effects and interactions independent
Yes, main effects and interaction tests in factorial ANOVA designs are always independent.