Statistical inference Flashcards
Goal of statistics
To be able to make inferences about a population, but you can never truly know about an entire population
Instead, you must use a sample, and ensure that the sample and inferences are actually generalizable to the population
Sampling variation
Refers to the inevitable differences between our population and our sample.
We want to minimize this as much as possible in order to have a more accurate sample
Have to think about getting best possible estimate in face of sampling variation
Sampling bias
refers to systematic forces with regard to sampling
Ex. only using university students, or people who pick up the phone/agree to do a study over the phone
Frequentist statistics
Frequentist statistics asks: how likely is it that this effect of the manipulation would have occurred if there was no effect in the actual population?
Ie. what is the chance this result was due to sample variation?
Is there another reason another reason this effect has arisen? Manipulation vs. random? What would happen if we did this study multiple times
Determining the answer to this question is the point of significance testing
What is more common Frequentist vs Bayesian stats
Frequentist statistics is more common: t-tests, p-values, null hypothesis
Sampling distribution
generally focuses on the shape of the charted data
Central limit theorem
states that for a large enough sample size, the mean of the sample will be…
Normally distributed
Equal to the population mean
Have equal variance to the population variance (if divided by sq root of sample size)
This is true regardless of the distribution of the actual population
Standard error
the standard deviation of the sampling error
Will get smaller as the sample size gets bigger, therefore can be used as a measure of sampling error variation/precision
Essentially, standard deviation is a measure of the spread of the data, standard error is a measure of the variation of the sampling error
We cannot actually know the true population distribution, but we can use CLT and standard error to make inferences
Confidence intervals:
Using CLT and standard error to infer if sample mean is true to population mean
Through sub-sampling, we are able to make inferences called confidence intervals
After getting the mean of multiple subsamples, you can see what percent of the means fall within one SD on either side of the mean
The confidence error tells us that 95% of repeated measures will fall within the SE of the sample mean
95% CI allows you to estimate that 95% of estimates lie within the specified SD
Bigger sample size equals smaller confidence intervals, because bigger sample size equals smaller SD, therefore the mean is more precise
Null hypothesis (h0)
assumes that results obtained are the result of random chance/sampling variation, and not an effect of the manipulation
Alternative hypothesis (h1)
assumes the results obtained were due to an effect of the manipulation.
p-value and null hypothesis
P-value refers to the probability that a test-statistic this extreme would be obtained given the null is true/by random chance (conditional probability)
If the p-value/the probability of receiving this result by chance is very small, you can reject null hypothesis and assume the result is the effect of the manipulation
I.e., if the p is very small, H0 is probably wrong, so we should reject it
P-value is industry standard/arbitrarily set at 0.05 (if smaller or equal too, usually reject null, if bigger, accept null
How to get p-value
favored tool for hypothesis testing
Essentially creates a null population distribution: what would we expect to see if this study was repeated and there is no real effect?
To get p-value, take standard error and calculate how many SE away this is the sample’s mean from the pop mean
Do this with z-score
Z-score = (Samp-mean-pop mean)/SE
The probability of observing a sample with our mean randomly is our p-value.
If smaller than 0.05, this can be considered statistically significant.
Type 1 error
False positive ie., reject the null despite there being no real effect
Alpha is the probability of a type 1 error
True positive is 1-beta
False positive is alpha
Type 2 error
False negative ie., accept the null despite there being a real effect, can combat through increasing power
Beta is the probability of a type 2 error
True negative is 1-alpha
False negative is beta