Trust Economics Week 9 Flashcards
This topic explores how much we should trust empirical results published in journals.
FACTS: replicability vs non-replicability
Replicability exists in sciences why we have law of physics etc.
Non-replicability in economics as studying animal behaviour (no heterogeneity when responding to stimuli)
What is the standard economic method of testing
Hypothesis testing i.e start with null.
E.g drug doesn’t cure cancer. (No relationship on cancer)
Reject null means drug does cure cancer. (There is a relationship)
P value
How unlikely the pattern in your data would have arisen, if null is true. (E.g drug does not cure cancer, but how surprised am I to see data that shows it does?) (level of confidence we can reject the null, smaller=better)
(Small p value=greater evidence null hypothesis is false, so drug does cure cancer)
2 problems with research practices harming credibility/trustworthiness of results
Publication bias
P-hacking
Publication bias
Publication likelihood greater with a smaller p-value
I.e smaller p-value = ones that have a significant effect, and reject the null.
P-hacking
Researchers make methodological choices in conducting their studies that tend to deliver lower P-values i.e significant effects. (Again no-one wants boring new with no relationship)
Remember P-hacking can still be accurate, just bias.
What is publication bias an example of?
A selection effect
Selection effect and example
What we see in newspapers/journals is not everything, but a filtered subset. Selection is not usually random, so needs to be considered!
e.g the WW2 Study by RAF of returning bombers and their pattern of bullet holes. TEACHES YOU TO ACKNOWLEDGE HOW SELECTION HAPPENED, AND THE SUBSEQUENT REASON FOR RESULTS
Why does publication bias arise
People want to see X is related Y, not X isn’t related to Y!
E.g people wanna see coke makes you bald, not coke doesn’t make you bald.
Opposed to nature of science, how does nature of statistics work? And examples
Enough studies enough times can create a purely spurious result
Jelly bean colour example, coke example.
Unlike where you apply heat to water, it always becomes steam.
Coke example explained
Questions men about their consumption for 15 drinks. Then whether bald or not.
Most drinks had nothing to do with hair loss, represented by blue dots.
Journalist only writes up on red dot (significant result)
No causation, but a positive correlation between coke o drinkers and being bald. (Maybe cos older men drinking more fizzy drinks!)
Examples of p-hacking decisions the researcher can make (3)
What data to collect
What sample
Defining variables
Example of p-hacking in hotel reviews
Hotel only request feedback for online review if they know their customers enjoyed their stay, hence why reviews are generally good on tripadvisor
Results can be probabilistic rather than deterministic- example
E.g giving employee of month might not motivate every worker, but a change in effort across a large sample/portion