Resampling and Open Science Flashcards

1
Q

What are resampling methods?

A

Sets confidence intervals and critical values of tests
Non-parametric
Relatively assumption free
Computationally demanding

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Bootstrapping vs Permutation

A

Both involve computing simulations
Bootstrapping samples with replacement
Permutation samples without replacement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Simulating Null Hypothesis Distribution Permutation Tests

A

common statistical question usually comparing two groups

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Null Hypothesis

A

States that the group means are equivalent

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Alternative Hypothesis

A

states that the groups means are not equivalent

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

P-value

A

Indicates probability of obtaining a difference in the mean of the two samples at least as extreme as observed at random, if the two samples did come from the same population

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Null Hypothesis Testing using Parametric Methods

A

Simulate or select theoretical null hypothesis sampling distribution
Determine where our observed test statistic lies within this distribution and the probability of it being observed if null hypothesis was true

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Null Hypothesis Testing using Permutation Tests

A
  1. calculate the real difference between the means of two groups
  2. simulate or select theoretical null hypothesis sampling distribution
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Simulating H0: Random Shufflying of Observations

A

Pool sample together then draw new groups from pooled sample
Repeat process many times ensuring it is not due to chance
Build a mean that should be true under the null hypothesis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Bootstrapping

A

Resampling method for assessing statistical accuracy of an estimate
Own sample and treat it as the entire population assessing it better
does not work for small samples
Typically used to estimate quantities associated with the sampling distribution of estimates
Original sample is resampled then drawn from which may be repeated multiple times creating a bootstrapping sample

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Basic Ideas of Bootstrapping

A

Treat particular sample as the entire population
Repeatedly sample with replacement to generate samples
Analyse to get estimate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How Science Should Work?

A

Thought to be a reliable way to answer questions about the world
Start with hypothesis
Collect data
Do statistics to test null hypothesis
Make conclusion based on the data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Reproducibility Crisis in Science

A

Findings do not replicate
Most significant effects should replicate
97% of 100 papers reported significant findings but only 37% were significant in the replication study
Data should reproduce as repeat experiments find significant effects replicating - rare

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Publication Bias

A

Not all research is equally publishable - editorial bias, incentivised, increased likelihood of false positives published
Wrong incentive structures - academic success tied to ‘significant’ results, publish or perish
Distorts meta-analyses - bad for estimating effect sizes
Not publishing negative results skew meta-analyese

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

P-Hacking

A

Actively searching for ‘something significant’ in data
Cherry-picking - experiments, subjects, stopping rules
Analysis of degrees of freedom
Variant of multiple comparisons problem
Focus on data giving significant findings and ignore those that do not inflate the chance of it occurring
Multiple comparisons unless controlled inflate probability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

HARKing

A

Hypothesis after results are known
Related to p-hacking
Hides the reality of multiple comparison problem

17
Q

Apophenia

A

Tendency to see patterns in random data

18
Q

Confirmation Bias

A

Tendency to focus on evidence that is in line with our expectations or favoured explanation

19
Q

Hindsight Bias

A

Tendency to see an event as having been predictable only after it has occurred

20
Q

Why do Findings Fail to Replicate?

A

When low-powered studies show significant effects, these will be overestimates

21
Q

Open Science - Open Hypothesis Testing and Analysis Decision Making

A

Pre-registration - publicly commit to your hypothesis and analysis pipeline before conducting study - treats p-hacking and HARKing
Registered report - coupled to publishing can also treat public bias

22
Q

Open Science - Open Analysis Tools and Data

A

Make all materials open access so other researchers can double check conclusions
Treats honest mistakes

23
Q

Open Science - Open Evaluation: Peer Review Published with Article

A

More information is always better

24
Q

Open Science - Open Access

A

Make the research outcomes (published articles) accessible to everyone