CHAPTER TWO Bayes in Science Flashcards

1
Q

What is the replication crisis in science?

A

A situation where many scientific studies are unable to be reproduced, raising questions about their validity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Who was Diederik Stapel?

A

A social psychologist known for committing fraud by fabricating data in his studies.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What was the significance of Daryl Bem’s study published in 2011?

A

It claimed to provide evidence for psychic powers through a priming experiment that defied traditional expectations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Define ‘priming’ in social psychology.

A

A technique where exposure to one stimulus influences a response to a subsequent stimulus.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What was the major twist in Daryl Bem’s priming experiment?

A

Subjects were primed after they had already made a choice, yet the priming still affected their responses.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What does a p-value of 0.01 indicate?

A

There is a 1% probability that the observed results would occur if the null hypothesis were true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

True or False: Daryl Bem believes his study results indicate that clairvoyance is real.

A

True.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What was the purpose of the ‘False Positive Psychology’ paper?

A

To demonstrate that standard statistical practices in science are flawed and can yield impossible results.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What was the unexpected finding in the ‘False Positive Psychology’ paper regarding age?

A

Listening to ‘When I’m Sixty-Four’ made participants appear nearly eighteen months younger.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Who is John Ioannidis?

A

A scientist who wrote about the prevalence of false findings in published research.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the misconception about p-values according to Dennis Lindley?

A

That a p-value of 0.05 indicates a 5% probability that the null hypothesis is true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Fill in the blank: The practice of analyzing data in multiple ways to find significant results is known as _______.

A

p-hacking.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is HARKing?

A

Hypothesizing After the Results are Known; formulating hypotheses based on observed data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the competitive reaction time task (CRTT) used to measure?

A

Aggression, particularly in research into the psychological effects of video games.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

True or False: The opponent in the CRTT is a real person.

A

False.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What drives the publish-or-perish culture in academia?

A

The need for researchers to publish regularly to secure tenure and career advancement.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What do most science journals prefer to publish?

A

Results that are interesting and novel, rather than all submitted studies.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What is the implication of publishing primarily ‘interesting and novel’ results in scientific journals?

A

It may lead to a bias where studies finding nothing significant are less likely to be published, skewing the literature towards positive results

This bias can create a false perception of the prevalence of certain phenomena, like psychic powers.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What threshold do many journals, especially in social sciences, use to determine if results are ‘interesting’?

A

p < 0.05

Results with p-values below this threshold are often published, while those above may be rejected.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

What is the potential issue with the publication of studies reporting on psychic powers?

A

If many studies find significant results (p < 0.05) while most do not, it can lead to a misleading conclusion that psychic powers are real

This situation can result in a majority of published studies supporting a claim that lacks sufficient evidence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What is p-hacking?

A

The practice of manipulating data to achieve statistically significant results

This can involve changing hypotheses or data analysis methods post hoc to obtain a desired p-value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Who is Brian Wansink and what was he known for?

A

A food scientist at Cornell University known for publishing numerous studies on eating behavior, some of which were later retracted due to scientific misconduct

His work included studies that were criticized for questionable research practices.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

What was the outcome of Wansink’s research practices?

A

Eighteen papers were retracted, seven received expressions of concern, and fifteen were corrected

Wansink resigned from Cornell in 2019 due to findings of scientific misconduct.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What did Brian Nosek’s Reproducibility Project investigate?

A

The ability to replicate results from one hundred psychology studies

The project highlighted the replication crisis in psychology, revealing that only 36 out of 100 studies could replicate their original findings.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

What was the average effect size found in Nosek’s replication study compared to the original studies?

A

The effect sizes of the replications were, on average, half the size of the originals

This discrepancy raised concerns about the reliability of published psychological research.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

What is a common misunderstanding about p-values among students and professionals?

A

Many believe that a p-value indicates the probability that the null hypothesis is true given the data

In reality, a p-value measures how likely the observed data would occur under a specific hypothesis.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

What is the fundamental difference between frequentist and Bayesian interpretations of statistical significance?

A

Frequentists assess evidence based on long-run probabilities, while Bayesians incorporate prior probabilities to assess the likelihood of a hypothesis being true

This distinction affects how researchers interpret p-values and statistical findings.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

Fill in the blank: A p-value of 0.02 indicates that there is a ______ chance of observing the data if the null hypothesis is true.

A

2%

However, this does not imply a 98% chance that the hypothesis is true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

True or False: A p-value of 0.05 means there is a one-in-twenty chance that the hypothesis is false.

A

False

The p-value does not provide the probability of the hypothesis being false; it indicates the likelihood of observing the data under the null hypothesis.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

What is a significant flaw in frequentist models according to the text?

A

They don’t take into account prior probabilities.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

What incentive do scientists have under frequentism when conducting experiments?

A

To come up with novel, surprising theories to get more buzz and notoriety.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

What does Bayesian statistics provide to scientists?

A

A vehicle for skepticism and a way to express disbelief in a theory.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

What does Aubrey Clayton suggest about hypotheses with low prior probabilities?

A

New data doesn’t significantly change skepticism towards them.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

What was the extraordinary finding at CERN in 2011?

A

Neutrons arrived at a detector 60 billionths of a second sooner than expected.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

What is the statistical significance of the CERN result?

A

A p-value of about 0.000000002.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

What fundamental axiom of relativity would be violated if the CERN finding were true?

A

Nothing can break the speed of light.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

What was the actual cause of the erroneous CERN result?

A

A fiber-optic cable in the clock system was not properly connected.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

What is Lakens’ position on publishing startling but low-prior results?

A

Such results shouldn’t be hidden just because of their low prior probabilities.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
39
Q

What did Karl Popper argue about scientific hypotheses?

A

You can never prove a scientific hypothesis; you can only disprove it.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
40
Q

What does Lakens label the belief that one can determine the probability of a theory being true?

A

The statistician’s fallacy.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
41
Q

What does Lakens prefer over using prior probabilities when evaluating data?

A

Letting the p-values stand for themselves.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
42
Q

What should scientists focus on to improve scientific findings according to Lakens?

A

Obtaining better data and raising the standard for statistical significance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
43
Q

What philosophical issue did David Hume raise?

A

The problem of induction.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
44
Q

What did Hume conclude about our expectation that the future will resemble the past?

A

It is based on custom and is ultimately unprovable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
45
Q

How did Karl Popper respond to the problem of induction?

A

He argued that science relies on falsification rather than confirmation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
46
Q

What example did Popper use to illustrate his point about scientific hypotheses?

A

The hypothesis ‘All swans are white.’

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
47
Q

According to Popper, what happens if you see a swan that isn’t white?

A

The statement ‘all swans are white’ cannot be true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
48
Q

What should scientists do when faced with a startling result?

A

Publish it despite low prior beliefs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
49
Q

What does Lakens think about the Bayesian approach in choosing research topics?

A

He uses it implicitly by selecting hypotheses he believes are likely to be true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
50
Q

Fill in the blank: Popper’s philosophy states that you can only _____ a scientific hypothesis.

A

disprove

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
51
Q

True or False: Lakens believes that knowing which theories are true is achievable.

A

False.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
52
Q

What is Aristotelian logic’s stance on inferring universal laws from individual examples?

A

You can’t infer a universal law from individual examples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
53
Q

What is a syllogism example demonstrating a logical fallacy regarding swans?

A

This is a swan, this swan is white, ergo all swans are white.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
54
Q

What does the statement ‘all swans are white’ imply if a non-white swan is observed?

A

The statement cannot be true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
55
Q

How does Popper believe science advances?

A

By falsifying false hypotheses, not by confirming true ones.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
56
Q

What method does Popper suggest for discovering regularities?

A

Trial and error, conjecture and refutation, or learning from mistakes.

57
Q

What term does Popper use to describe a theory that has stood up to severe testing?

A

Corroborated.

58
Q

What is the difference Popper suggests between two hypotheses?

A

One has been severely tested, and one has not.

59
Q

What is the critique of Popper’s model regarding the testing of theories?

A

It suggests that severely tested theories are not necessarily more likely to be true.

60
Q

What do Bayesians like Wagenmakers argue against Popper’s falsification model?

A

Most scientific hypotheses are not straightforwardly falsifiable by a single counterexample.

61
Q

In the context of acetaminophen, what do we claim about its efficacy?

A

Statistically, you are more likely to recover quickly from a headache if you take acetaminophen.

62
Q

What is the significance of a p-value below 0.05 according to Fisher?

A

You tentatively behave as though a hypothesis is true.

63
Q

What do frequentists reject that Bayesians encode as numbers?

A

Prior knowledge.

64
Q

What fundamental question does Wagenmakers argue researchers are interested in?

A

How likely is this hypothesis to be true, given the data we’ve seen?

65
Q

What did the authors of papers in Nature Human Behaviour say about their claims after seeing the data?

A

The data had made the claim more plausible than it was before.

66
Q

What is a basic advantage that Bayesians have over frequentists?

A

They don’t have to leave data on the table.

67
Q

What happens to data from previous studies in frequentist methods?

A

It gets forgotten with each new study.

68
Q

What is the expected outcome of running an experiment twenty times with a p-value of 0.05?

A

You’d expect to see a significant result once.

69
Q

What can happen if a researcher checks their data after each new entry in an experiment?

A

They can easily find statistically significant results even when there is no real effect.

70
Q

How do Bayesians handle new data in relation to prior beliefs?

A

Each new data point moves their opinion less.

71
Q

What is the implication of optional stopping for Bayesian analysis?

A

It can yield the same probabilities as waiting for all results.

72
Q

What is a limitation of frequentist statistics regarding hypotheses?

A

They only reject or accept the null hypothesis without giving degrees of belief.

73
Q

What is a common assumption about the null hypothesis in studies involving human populations?

A

The null hypothesis is always ultimately false.

74
Q

What is a null hypothesis?

A

A hypothesis that assumes no effect or no difference exists in the population being studied.

75
Q

Is the null hypothesis always true when studying human populations?

A

No, the null hypothesis is always ultimately false in human populations.

76
Q

What happens when a large enough sample size is used in hypothesis testing?

A

You will likely be able to reject the null hypothesis.

77
Q

What did David Bakan find in his 1968 study involving 60,000 subjects?

A

Every test he conducted came out significant, showing small but significant differences based on arbitrary criteria.

78
Q

What was Paul Meehl’s observation regarding his survey of Minnesota high school students?

A

92 percent of the possible combinations of responses showed statistically significant correlations.

79
Q

What is a key distinction between frequentist and Bayesian statistics?

A

Frequentist statistics requires deciding to reject or not reject the null hypothesis, while Bayesian statistics estimates the size of the effect and provides a probability distribution.

80
Q

What does a probability distribution represent?

A

A graph of the possible outcomes of an event.

81
Q

What is a prior distribution in Bayesian statistics?

A

The best estimate of the size of an effect before seeing the data.

82
Q

What is a likelihood curve in Bayesian analysis?

A

The distribution of new data around an average after observing the data.

83
Q

What is a posterior distribution?

A

The curve that combines the prior distribution and the likelihood to represent updated beliefs after seeing new data.

84
Q

What is the implication of a tall and narrow prior curve in Bayesian analysis?

A

It indicates strong confidence in the prior estimate.

85
Q

What does the term ‘Lindley’s paradox’ refer to?

A

The idea that a p-value of 0.05 can be evidence against a hypothesis under certain conditions.

86
Q

What does a p-value of 0.05 signify?

A

It indicates a one-in-twenty chance of seeing results as extreme as those observed under the null hypothesis.

87
Q

How can p-values be misleading in hypothesis testing?

A

A statistically significant p-value does not provide evidence for the null hypothesis, only that the data is surprising under that hypothesis.

88
Q

What happens to the distribution of p-values if there is a real effect?

A

The p-values cluster closely around zero.

89
Q

What does Kevin McConway suggest about p-values?

A

They are useful for indicating surprise under the null hypothesis but do not indicate the likelihood of the null hypothesis given the data.

90
Q

Fill in the blank: The Bayesian approach allows you to update your _____ with new information.

A

probability distribution

91
Q

True or False: Frequentist analysis can solve all problems in modern science.

92
Q

What does a p-value of 0.05 indicate?

A

It indicates that the data is surprising given the null hypothesis.

It does not provide information about the likelihood of the null hypothesis being true given the data.

93
Q

What is a common misconception among researchers regarding p-values?

A

Many researchers assume that a statistically significant result confirms their hypothesis.

This is not the correct interpretation of what a p-value represents.

94
Q

Why is a p-value threshold of 0.05 considered weak evidence?

A

Because it can lead to false confidence in results that may not be statistically significant in practical terms.

For example, rolling two sixes on loaded dice can yield a p-value of 0.028.

95
Q

What did a group of scientists propose regarding the standard level of statistical significance in 2017?

A

They proposed to lower the threshold for statistical significance to 0.005.

This would reduce the number of situations where Lindley’s paradox is relevant.

96
Q

What is Lindley’s paradox?

A

It refers to the situation where a statistical hypothesis test can provide contradictory conclusions depending on the p-value threshold used.

97
Q

What is Bayesianism concerned with in statistics?

A

It focuses on updating beliefs and assessing probabilities based on prior knowledge and new evidence.

This contrasts with frequentist approaches that rely solely on the data at hand.

98
Q

What are the two types of uncertainty discussed?

A
  • Aleatory uncertainty
  • Epistemic uncertainty
99
Q

What is aleatory uncertainty?

A

It is the uncertainty in an unknowable future, such as the outcome of a coin flip or a plane landing.

The term comes from the Latin word ‘alea’, meaning ‘a die’.

100
Q

What is epistemic uncertainty?

A

It is the uncertainty regarding knowledge of a past event or fact, such as whether a coin landed heads or tails after being flipped.

101
Q

How is probability viewed in a Bayesian framework?

A

Probability is seen as a measure of subjective belief based on available information.

102
Q

What is a uniform prior?

A

It is a prior probability distribution that assigns equal probability to all possible outcomes, indicating total ignorance.

Frequentists often implicitly assume a uniform prior.

103
Q

What problem is associated with uniform priors?

A

They can lead to paradoxes where total ignorance in one sense results in strong prior beliefs in another sense.

104
Q

What did Harold Jeffreys suggest regarding prior distributions?

A

He suggested a U-shaped prior distribution, concentrating probability mass heavily at the extremes.

105
Q

What is the significance of incorporating expert opinion into priors?

A

It allows for more informed decision-making rather than relying solely on uniform priors or data.

Experts can provide valuable insights that should be integrated into the prior probabilities.

106
Q

How can the robustness of conclusions be checked in Bayesian analysis?

A

By testing different prior distributions to see if conclusions remain consistent.

107
Q

What is meta-analysis in the context of Bayesian statistics?

A

It is the process of combining data from multiple studies to create a consensus, which is integral to Bayesian analysis.

108
Q

True or False: Frequentists believe in using prior distributions based on previous studies.

A

False.

Frequentists typically do not use prior distributions in their analysis.

109
Q

What is a thought experiment to determine if one is a Bayesian or a frequentist?

A

Flipping a coin and assessing the probability of it being heads without looking at the result.

110
Q

What does it mean if someone answers ‘50 percent’ regarding the probability of the flipped coin being heads?

A

They are thinking in a Bayesian way, considering their subjective beliefs and available information.

111
Q

What is the implication of believing that the probability of a past event is either 0 or 100 percent?

A

It indicates a frequentist perspective, focusing on the actual outcome rather than subjective probability.

112
Q

What is a meta-analysis?

A

A method where scientists pool all studies on a subject to create a consensus by combining data like p-values and effect sizes.

113
Q

What is a meta-analytic prior?

A

A current standard way of creating a prior distribution from existing data, which incorporates meta-analysis.

114
Q

Why is using a proper, informative prior important in Bayesian statistics?

A

Not using existing data leads to less certain conclusions and inefficient use of new data.

115
Q

What is a potential problem when using priors in Bayesian analysis?

A

Strange priors can skew results, especially if they misrepresent control group expectations.

116
Q

How can one avoid skewing results with priors?

A

By using a mix of distributions for the prior, which down-weights historical information if there’s a significant difference.

117
Q

What does the term ‘objective’ refer to in Bayesianism?

A

It describes a school of thought within Bayesian statistics that aims for neutrality in prior selection.

118
Q

What analogy does Sophie Carr use to explain Bayesian and frequentist statistics?

A

She compares them to rugby league and rugby union, highlighting that both have pros and cons.

119
Q

What is one criticism of frequentist methods mentioned in the text?

A

The p = 0.05 threshold is considered weak and could be improved.

120
Q

What is a Registered Report?

A

A model where journals publish papers based on the strength of their methods before data collection.

121
Q

What is the Octopus program?

A

A free repository for hypotheses, data, code, and methods, designed to change the incentive structure in research.

122
Q

What is one reason why the debate between Bayesian and frequentist statistics persists?

A

Bayesian methods are becoming more common, but frequentist techniques remain the standard in scientific inquiry.

123
Q

What is the Lindley paradox?

A

Under frequentist analysis, a statistically significant result can actually be evidence against the hypothesis.

124
Q

How does Bayesian analysis address the issues of the replication crisis?

A

It forces comparison of likelihoods between competing hypotheses, making it harder to misinterpret results.

125
Q

What advantage does Bayesian analysis have in vaccine studies?

A

It allows for the incorporation of prior data to reach significance thresholds more quickly.

126
Q

What does Eric-Jan Wagenmakers say about the aesthetic of Bayesianism?

A

Bayesianism is aesthetically more pleasing due to its coherence and lack of internal inconsistencies.

127
Q

What is utility in the context of decision theory?

A

It refers to the satisfaction or benefit derived from choosing one option over another.

128
Q

Fill in the blank: Using a proper, informative prior in Bayesian analysis prevents _______.

A

leaving money on the table.

129
Q

True or False: The frequentist method is the only standard for scientific inquiry.

130
Q

What is a downside of the academic journal model according to Marcus Munafo?

A

It is outdated and does not reflect the complexity of modern research.

131
Q

What is one of the main problems with p-values in frequentist statistics?

A

They can lead to p-hacking and do not account for prior probabilities.

132
Q

What is the underlying system for all decision-making according to the text?

A

Bayes

Bayes refers to Bayesian probability, a method of statistical inference.

133
Q

What is the title of the recommended Coursera course?

A

Improving Your Statistical Inferences

The course is by Lakens and is available for free.

134
Q

True or False: Frequentists cannot estimate effect size within their framework.

A

False

Frequentists can estimate effect size and perform equivalence testing.

135
Q

What does equivalence testing allow you to determine?

A

Whether an apparent effect is big enough to pay attention to

Equivalence testing is a method used to assess if the effect size is practically significant.

136
Q

Fill in the blank: Rejecting or accepting the null hypothesis isn’t _______.

A

final

This indicates that statistical conclusions can be revisited.

137
Q

What is a criticism mentioned regarding the terminology used by statisticians?

A

They use terms that have different technical meanings from everyday language

This creates confusion, similar to the term ‘significance’.

138
Q

What is the implication of the statement about frequentists not being idiots?

A

Frequentists have considered alternatives and developed methods within their framework

This acknowledges the validity of frequentist approaches.

139
Q

What is a key difference between Bayesian and frequentist approaches as suggested in the text?

A

Bayesian methods incorporate effect size estimation into the system, while frequentist methods do so ad hoc

This highlights a structural difference in how each approach handles statistical inference.