Trust Economics Week 9 Flashcards

1
Q

This topic explores how much we should trust empirical results published in journals.

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

FACTS: replicability vs non-replicability
Replicability exists in sciences why we have law of physics etc.

Non-replicability in economics as studying animal behaviour (no heterogeneity when responding to stimuli)

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the standard economic method of testing

A

Hypothesis testing i.e start with null.

E.g drug doesn’t cure cancer. (No relationship on cancer)

Reject null means drug does cure cancer. (There is a relationship)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

P value

A

How unlikely the pattern in your data would have arisen, if null is true. (E.g drug does not cure cancer, but how surprised am I to see data that shows it does?) (level of confidence we can reject the null, smaller=better)

(Small p value=greater evidence null hypothesis is false, so drug does cure cancer)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

2 problems with research practices harming credibility/trustworthiness of results

A

Publication bias

P-hacking

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Publication bias

A

Publication likelihood greater with a smaller p-value
I.e smaller p-value = ones that have a significant effect, and reject the null.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

P-hacking

A

Researchers make methodological choices in conducting their studies that tend to deliver lower P-values i.e significant effects. (Again no-one wants boring new with no relationship)

Remember P-hacking can still be accurate, just bias.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is publication bias an example of?

A

A selection effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Selection effect and example

A

What we see in newspapers/journals is not everything, but a filtered subset. Selection is not usually random, so needs to be considered!

e.g the WW2 Study by RAF of returning bombers and their pattern of bullet holes. TEACHES YOU TO ACKNOWLEDGE HOW SELECTION HAPPENED, AND THE SUBSEQUENT REASON FOR RESULTS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Why does publication bias arise

A

People want to see X is related Y, not X isn’t related to Y!

E.g people wanna see coke makes you bald, not coke doesn’t make you bald.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Opposed to nature of science, how does nature of statistics work? And examples

A

Enough studies enough times can create a purely spurious result

Jelly bean colour example, coke example.

Unlike where you apply heat to water, it always becomes steam.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Coke example explained

A

Questions men about their consumption for 15 drinks. Then whether bald or not.

Most drinks had nothing to do with hair loss, represented by blue dots.

Journalist only writes up on red dot (significant result)
No causation, but a positive correlation between coke o drinkers and being bald. (Maybe cos older men drinking more fizzy drinks!)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Examples of p-hacking decisions the researcher can make (3)

A

What data to collect
What sample
Defining variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Example of p-hacking in hotel reviews

A

Hotel only request feedback for online review if they know their customers enjoyed their stay, hence why reviews are generally good on tripadvisor

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Results can be probabilistic rather than deterministic- example

A

E.g giving employee of month might not motivate every worker, but a change in effort across a large sample/portion

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

We need to consider scale of these problems.

A
17
Q

Stylised example without publication bias or p-hacking

A

Consider 1000 hypotheses, 50 true. We don’t know which. So experiment done on all 1000.

Experiment finds significant result if actually true=0.8 (“Power” of test)
Experiment finds significant result if actually false=0.05 (arbitrarily p-value cut-off)
Seeing a positive result, the chance of underlying hypothesis being true is

(0.8 x 50) / {(0.8 x 50) + (0.05 x 950)}= 0.46

18
Q

Meaning of this

A

If we see a paper at a 5% significant that the result is true, the probabitly of actually being true is 0.46

LESS THAN HALF!

This example shows the extent of the problem, REMEMBER THIS EXPERIMENT HASN’T EVEN HAD PUBLICATION BIAS OR P-HACKING, SO IN THE REAL WORLD EFFECT EVEN WORSE

19
Q

So what should we consider more (2)

A

Theory along with results to support the findings

High powered studies, more compelling with a bigger sample.

20
Q

Marathon analogy

A

We would assume a small amount of people at the beginning, a lot in middle range and few at ending- a normal distribution.

In reality there are multiple spikes; people push to get 2:59 rather than 3:01 etc. so these cause the spikes with more people finishing at these times

21
Q

Brodeur et al

A

Collected 13440 p-values to see their z-statistics distribution when unmanipulated!

Spikes around critical values (like the marathon example).

Show a major filtration process, meaning we don’t see certain results! (Ones where fail to reject i.e the boring ones)

22
Q

So first way to collect evidence is to collect P-values and see the distribution (marathon and Brodeur).

Second way to see more evidence, and example

A

Try to replicate studies

If we collect a bunch of studies with p<0.05, this means more than 95% should replicate!

Nosek- replicated 100 research findings, only 39 could be reproduced. (39 IS NOT 95!!!)

23
Q

So, 2 ways to show evidence and increase accuracy

A

Collect p-values and see their distribution

Replicate studies

24
Q

Solutions

A

Change publication practices to allow null results to be more easily published.

Encourage replications via data sharing, journals publishing replications, university recognition etc.

Make replications easier e.g via relaxing data-sharing requirements: “open science” movement

Require “pre-registration” or “pre-analysis plans” to address p-hacking

25
Q

Change publication practices to allow null results to be more easily published. Example

A

Journal of Development Economics receive experiments without results. So reviewer and editor only decide to publish on the basis of the quality of the methodology, rather than the answer!

26
Q

Require “pre-registration” or “pre-analysis plans” to address p-hacking

A

Make experimenters state their choices and state what they’re going to do, before collecting their data.

So reviewers know how these results arised prior.