Midterm Flashcards
The General Linear Model Goal
try to account for as much variability in a dependent variable as possible
the Y variable must be continuous (interval or ratio scale)
can use one or multiple IV’s to account for the variance in the DV
Categorical Variables
Binary Variable- 2 distinct categories
Nominal Variables- more than 2 distinct categories
Ordinal Variables- more than 2 distinct categories which go in a logical order
Continuous Variables
entities get a distinct score on a scale
Interval Variable: equal intervals on the variable represent equal differences in the property being measured
Ratio Variable: same as interval, but the ratios of scores are important and 0 is meaningful
Measuring Effect Size
R2 (Coefficient of Determination)
how much variance in the outcome variable is accounted for by the IV
We test whether the effect size is significant by the F and t statistics
The Statistical Model
We want to try and theoretically reflect real world phenomena
we want to be 95% sure that our findings are due to our model (only 5% chance it happened by chance)
Sum of Squares Total
The total variability between scores and mean
The sum of each score minus the mean squared
Sums of Squares Error
Deviance between the model and each person’s predicted score
The sum of each score minus the predictor score squared
Sums of Squares Model
deviance between the mean and the model
The sum of The mean subtract each person’s predicted score squared
Mean Square
The average of the sum of squares
the SS divided by the associated degrees of freedom
Degrees of Freedom
the wiggle room in the data set
because our mean must stay constant, all of our scores can be anything except our last score, which must bring us to our mean
Central Limit Theorem
if there are 30 or more participants in a study, a normal distribution will begin to emerge
Confidence Intervals
Describes the upper and lower bounds of a score
We want to be 95% sure the score will land in the confidence intervals
Types of Hypotheses
Null Hypothesis- we assume there is no effect of the IV on the DV
Alternative Hypothesis- that there is an effect of the IV on the DV
We assume the null hypothesis until shown otherwise
One and Two Tailed Error
One Tailed- probability only goes one way (.05 on either the positive or negative side)
Two Tailed- probability is taken on both sides, .025 on each side
Type i error
when we believe there is a genuine effect and there is not
probability of this happening is measured at the alpha level (usually .05)
Type ii error
when we believe there is no effect and there is an effect
the probability of this happening is measured at the beta level (usually .2)
Confidence Intervals and Statistical Significance
if CI’s overlap, generally the findings are not significant
As sample size increases, CI’s decrease and we are more likely to find a significant result
Misconceptions about the P Value
- Significant result DOES NOT mean it is important
- A Non-significant hypothesis DOES NOT mean there is no effect, only that it is not big enough to be found
- A significant result DOES NOT mean the Ho is false
Problems with NHST
All or nothing thinking (that significance is everything; instead, we can also look at effect size)
Reliant on sample size
Wider Science Problems with NHST
Incentive structure- you are more likely to exceed in the research field if your findings are significant
Researcher Degrees of Freedom- a researcher’s decisions can change the P value and make it significant
P-Hacking (changing certain numbers or methods after the fact to make your P significant)
Harking- finding a significant result in your data you weren’t studying and then changing your hypothesis to match
Avoiding the Wider Science Problems
Open science- movement to make the process, data, and outcomes of research freely available
Pre-registering research- receiving feedback and promises of publishing by preregistering with a journal; ensures less competition
Effect Sizes
standardized measures o the size of an effect which can be compared across studies
not as reliant on sample size as p
Cohen’s d, Pearson’s r, and odds ratio are all examples
r, Correlation Coefficient
a good measure when group size is the same
A positive correlation suggests that the values increase or decrease together
A negative correlation suggests that as one increases, the other decreases
Effect Sizes of Pearson’s r
r=.1 (S)
r=.3 (M)
r=.5 (L)