L6 Ch5 Null Hypothesis Significance Testing Flashcards

1
Q

Disclaimer

A

There are quite a few of repetitions in these flashcards, and the ones from the book are not integrated but are in a separate second part
I am very sorry about this but I don’t have time to make them look nice, so I hope that at least they are clear enough to study

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

sampling distribution

A
  • distribution of means (usually)
  • how it would look if we were to repeat distributions over and over again
  • relates to null-hypothesis and alternative hypothesis (used to understand and interpret papers, in our case)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Fischer

A
  • inventer of the p-value & null-hypothesis
  • experiment with lady tasting milk or tea first
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Neyman-Pearson

A
  • inventors of alternative hypothesis
  • null-hypothesis and alternative hypothesis combined in one paradigm with p-value
    > tricky to specify what an alternative hypothesis is
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Standard Error

A

= variability in sampling distribution (variability that you can expect when repeating the experiment)
- SE high if lots of variability in variable
- SE low if high sample size
> high sample size → low variability → low SE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Frequentist probability

A
  • considers p-value and sampling distribution
  • computes objective probability of an event
  • relative frequency (outcomes of event) in the long run (over same test done multiple times)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

how can confidence intervals be interpreted?

A
  • compute CI for 100 samples, and create sampling distribution for said samples
  • a CI of 95% means that 95 out of the 100 CIs for the samples will contain the population mean
    ~ single CI either contains the true value or it doesn’t
    ~ wider or narrower based on how certain we are of the inference
    ~ much better than using point estimate
    (see picture 2)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

how can confidence intervals be calculated?

A
  • lowerbound: mean - 1.96 x SE
  • upperbound: mean + 1.96 x SE
    (picture 1)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the SE used for? How?

A
  • parameter estimation (for population)
    > through confindence intervals (usually 95%)
    ~ higher SE → higher variability → broader CI (to reach 95% confidence)
    ~ lower SE → lower variability → narrower CI (to reach 95% confidence)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

how can SE be calculated?

A

standard deviation / square root of sample size

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

sampling distributions under Ha

A
  • different than under H0
  • e.g. skewed
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

“R”
- what is it?
- R vs Excel
- in exam

A
  • can be used as simple calculator
  • much more extensive than excel
  • open source (important as science should be open)
    > primarly used as calculator (no extensive programming)
    > data simulation
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Binomial sampling distribution under H0
- how to compute it in R

A

“if probability of heads is 0.5, what is the probability of getting 8/10 heads?”
- remember to run all the lines!
1. n <- 10 (sample size)
2. k <- 0:n (discrete probability space)
> this means that k is equal to the number 0 to n (10)
3. p <- .5 (probability of head)
4. coin <- 0:1
5. permutations <- factorial(n) / ( factorial(k) * factorial(n-k) )

  • “barplot” function → give values of probabilities to function, and it constructs the plot
    (picture 3)
    ! look at WAs for representations of how R will be in the exam
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Type I error

A
  • reject null hypothesis when it is true
  • “false positive”
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

what are the possible outcomes if we make a decision in frequentist framework?

A

(see picture 4)
- rows: do we (not) reject the H0?
- columns: is the H0 actually true/false?
- two squares per correct or incorrect decision (type I or type II error)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Type II error

A
  • not reject null hypothesis when it is false
17
Q

how strict do we want to be when evaluating the H0?

A
  • decide on alpha level (usually 0.05
  • if p-value is below alpha level, we reject null hypothesis
    = type I error in 5% of the cases
    !! alpha is type I error rate
18
Q

effect sizes

A
  • size of effect that we are looking for (e.g. size of correlation)
  • plays a role in how much power our statistical procedure has
  • standardized (divided by st.dev.)
19
Q

how do we use the sampling distribution in regards to alpha?

A
  • we mark areas in sampling distribution that constitute an extreme enough observation that makes us reject H0
  • with extreme observations we reject H0 (picture 5)
20
Q

“power” of the analysis

A
  • reject H0 when it is in fact false (correct decision)
  • power is conditional probability of rejecting H0 when false
  • it’s a function of sample size
  • (alpha is the conditional probability of H0 being true)
21
Q

how are effect sizes and power related?

A

the bigger the effect size, the more likely we are to have a higher power

22
Q

what is the probability of not rejecting the H0 when it is true?

A
  • 1-alpha
  • “true negative”
    (see picture 6)
23
Q

what is the sum of the power and the true negative?

A
  • 1
  • they are conditional probabilities
24
Q

Beta

A
  • opposite of power
  • incorrectly decide to not reject the H0 when it is false
  • Type II error
  • “false negative”
25
Q

how does changing the value of alpha affect the evaluation of the H0?

A
  • lower alpha
    → harder to reject H0
    → less type I error
  • higher alpha
    → easy to reject H0
    → less type II error
    ! alpha used to establish critical region
26
Q

how do we calculate “power” in a sampling distribution?

A
  • power: rejecting H0 when not true
  • look at sampling distribution under Ha
    → there are many versions, but in the example of the dice, we could set the mean at 0.8
    (see picture 7)
  • with new distribution, what is probability of rejecting H0 now? (what is our power?)
27
Q

summary of this procedure

A
  • we decide when to reject the H0, based on the sampling distribution under the H0
  • look at sampling distribution + alpha level = reject H0?
  • then change to Ha distribution while keeping the “red regions” the same
  • now: what is the probability of rejecting H0 when Ha is true? (power)
    → power is the sum of probabilities in red

!!! to compute power, we look at sampling distributions under the Ha

28
Q

what is the interplay between alpha and power?

A
  • if low alpha level, we are stricter when rejecting H0
    → power decreases as well
  • balance between power and type I error rate
29
Q

effect size vs Ha

A
  • effect size: p of heads
  • increase effect size (further from 0.5)
  • more extreme values become more likely → more likely to reject null hypothesis
    ! conditioning on the effect size
30
Q

what determines power level?

A
  • alpha (lower alpha → lower power)
  • effect size (higher e.s. → higher power)
    > greater effect → more likely to reject H0 → more power
31
Q

From the book

A
32
Q

how can you distinguish a frequency plot from a histogram?

A
  • frequency plot has small gaps between the columns
33
Q

what is determined by the length of the whiskers?

A
  • if whiskers have same length, then distribution is symmetrical
  • if the top of bottom whisker is much longer than the opposite, then the distribution is asymmetrical
33
Q

how can you compare the relative frequencies of scores across groups?

A

Under frequency plots:
- stack: shows the bars of each group stacked on top of each other
- identity: displays each overlapping bars, with a certain level of transparecy
- dodge: places the bars side by side within each bin
(see picture 8 & 9)

34
Q

Boxplot

A
  • center: median
  • sides: interquartile range
  • violin element: includes density distribution of the data
    ~ using split variable, you can visualize the group difference
    (see picture 10)
34
Q

how can we summarize the relationship between two variables?

A
  • through a regression line
  • “correlation plots”
35
Q

Raincloud plots

A
  • display individual data points,boxplots and distribution of data points
    (see picture 11, 12 & 13)
36
Q

Gigamega mastermind mind-map of plots

A

picture 14