Advanced Data Design & Analysis Flashcards
Define and describe variance
For estimating population variance:
* Person score-population mean/population is unbiased
○ Makes fewer extreme estimates
○ Unrealistic though because it’s rare that we know the population mean
* (Person score-sample mean)^2/population is biased
* (Population score-sample mean)^2/population-1 is unbiased
○ This is unbiased because the sample mean is not the same as the population mean (used in the first option), the sample scores are going to be closer to the sample mean than the population mean, so the squared deviation terms are going to tend to be underestimates. Therefore, (n-1) corrects for that by making the result slightly bigger (which has a larger proportional effect when the n is small - because when the sample is larger it will be a closer estimation to the population mean)
A sampling distribution is a sample of difference variances estimated by the formulas, from it you can see how biased/unbiased a predictor is if you know the actual population variance. You can get a more accurate prediction because it creates an average of many estimates
What is power?
To say that a test has 80% power means it has 80% chance of rejecting the null hypothesis given that the null hypothesis is false, and given
* A particular sample size
* Particular effect size
* Particular alpha level (often .05 probability of rejecting the null hypothesis)
* Other considerations, including those related to whether the assumptions of the test are satisfied
Reflections on power:
* We don’t have great intuitions about sample size as it relates to power.
○ Our intuitions may have been warped by seeing psychology journals reporting findings with very small sample sizes but statistically significant results
§ Example of publication bias (when non-significant studies tend to get chucked out
Explain Statistical inference issues
-you cannot prove or disprove theories
-They provide probabilistic information and at most can corroborate theories
-significance tests do not allow probabilities to be assigned to any particular hypothesis
-with an alpha level of .05, we should reject the true null hypothesis 5% of the time (kind of like a margin of error). However, that is a global error rate, and doesn’t tell us the probability of a local mistake
Explain P-Values
p is the probability of getting our observed result, or a more extreme result, if the null hypothesis is true
Explain confidence intervals
- General formula for confidence intervals for the population mean is M +/- Margin of error
- If you’re randomly sampling from a normally distributed population:
- Issue is it is VERY rare to know the population standard deviation
- If you don’t know the standard deviation you will base it off a t-distribution rather than a normal distribution
- Cut-offs will be different, and with larger sample sizes it will be more similar to a normal distribution
- But don’t just use the cut-offs for a normal distribution because they won’t be the same
INTERPRETING CONFIDENCE INTERVALS
* If we ran many studies, 95% of the intervals would contain the population mean
* We don’t know whether this particular interval does or doesn’t
What is the difference between p-values and confidence intervals?
- Advantages of confidence intervals
○ They give a set of rules that, if they were your null, would have been rejected
○ They tell you about the precision of your estimate - An advantage of p-values
○ The give a clearer indication of the evidence against the null hypothesis
Describe the replication crisis and 5 suggested reasons for it
The ‘file drawer’ problem
* The bias introduced in the scientific literature by selective publication - chiefly by a tendency to publish positive results, but not to publish negative or nonconfirmatory results
Gelman (2016) mentions five reason:
* Sophistication
○ Since psych was focussing on more sophisticated concepts than other disciplines, it made it more open to criticism
* Openness
○ Has culture in which data sharing is very common - easier to find mistakes
* Overconfidence deriving from research design
○ Researchers may feel that they can’t go wrong using simple textbook methods, and their p-values will be ok
* Involvement of some prominent academics
○ More of its leading figures have been dragged into the replication crisis than other disciplines
* The general interest of psychology
○ Methods are very accessible so more people are willing to find mistakes
What are some routes to the replication crisis
- Outright fraud (rare)
- P-hacking, data dredging, data snooping, fishing expeditions (rarer than is commonly believed)
○ Looking for what people want to find
○ Sifting through the data - The garden of forking paths (more common than generally realised)
○ Only run the experiment that seems like it might be significant based on the data (looking at data before doing the tests, if data was different, might have done other tests)
○ Solution could be to make requirements to preregister the hypotheses and methods
What are some typical experimental designs?
Between-subjects design
* Different participants contribute data to each level of the IV
* But differences might be due to differences between participants
○ Random assignment can help reduce this
○ Or matching - balance out the conditions
Within-subjects design
* Each participant gets exposed to each level of the IV
* Major concern with this is sequencing effects (each previous level having an influence on the next level)
Single factor design
* Only one IV
* Can have two levels (each placebo and treatment), or more than two levels (eg placebo, treatment level 1, treatment level 2)
Factorial designs
* More than one IV
* Analysed with two-way ANOVA
* Interaction effects
Explain correlational research and its uses
- Investigates the relationships between two (usually continuour) variables - manipulate on variable to observe the effect on the other
- This type of research is often linked with th concepts of correlation and regression
○ Regression
§ Predicting a variable from other variables in a regression model - Designs are useful when
○ Experiments cannot be carried out for ethical reasons
○ Ecological validity is a priority
- This type of research is often linked with th concepts of correlation and regression
While correlation and regression are associated with correlational research, ANOVA and t-tests are associated with experimental research
This distinction is bogus, can use any of them with another research design
What is a quasi-experimental design?
○ Groups occur naturally in the world; cannot be random assignment to groups
§ Eg comparing men and women
○ Often used in program evaluation
§ Provides empirical data about effectiveness of government and other programs
What are some issues with null hypothesis significance testing
- If power is low, there may be a difference, but you don’t see it
- If power is very high (if you have a very large sample size you will likely have high power), even small differences can seem significant
Explain empiricism
out about the world through eveidence
○ Could be considered as:
§ Observation = truth + error
§ Observation = theory + error
§ Observation = model + error
○ We need a theory of error to fit our models
○ Classical methods in statistics tend to assume errors are normally distributed
§ Gauss was first to fullt conceptualise normal distribution
Why do we use linear models
○ Easy to fit
○ Commonly used
○ Lots of practical application (prediction, description)
○ Provide a descriptive model that is very flexible
○ Have assumptions that are broadly reasonable
What is the theory of error?
§ Often assume that the ‘error’ is normally distributed with zero mean
□ A theory of error
® It is the error term that requires statistical techniques
® The real question - eg relationship between age and IQ isn’t statistical at all
Why do we assume normal errors?
○ Two broad categories of justification for building models around the assumption of normal errors
§ Ontological
□ Study of nature of being
□ Normal distributions occur a lot in the world so let’s build model around them
□ Any process that sums together the result of random fluctuations has a good chance of somewhat resembling normal distributions
□ Sampling distributions and many statistics tend towards normality as sample size increases
§ Epistemelogical □ Normal distributions represent a state of our knowledge (more like ignorance) □ Don't contain any info about the underlying process, except its mean and variance □ Should still be interested in underlying process, but when we don't know anything about it, it's best to take in as few assumptions as possible
Explain common model assumptions
○ Validity
§ Ensure data is relevant to the research question
○ Representativeness
§ Want sample to represent population as well as possible
○ Additivity and linearity
§ Most important mathematical assumption
§ Want the non-error part of function to be a linear model
○ Independence of errors
○ Equal variance of errors
§ Also referred to as homogeneity/homoscedasticity
○ Normality of errors
Explain moderation vs mediation
- Moderation
○ Situations where the relationship between two variables depends on another variable- Mediation
○ Indirectly through another variable
○ Inherently causal
○ X causes M causes Y
○ Maybe X also causes Y, but it doesn’t have to
- Mediation
Explain mediation effects
- Mediating variables transmit the effect of an independent variable on a dependent variable
- Mediation is important in many psychological studies
- It is the process whereby one variable acts on another through an intervening (mediating) variable
○ Eg Theory of Reasoned Action
§ Attitudes cause
intentions, which cause behaviour - Simplest mediation model contains three variables
○ Predictor variable X
○ Mediating variable M
○ Outcome variable Y
○ Causal model
What is the causal steps approach to mediation
○ Based on the above regression equations and involves 4 requirements
§ X directly predicts Y (ie coefficient c is significant
§ If c is significant, X directly predicts M (ie coefficient a is signficiant)
§ M directly predicts Y (coefficient b is significant)
§ When both X and M predict the Y, the effect of X is either:
□ Reduced (coefficient c’ is smaller than c, though both remain significant), and there is partial mediation, or
□ Eliminated (ie coefficient c’ is not significant) - then there is full mediation
○ If any of the four requirements are not met stop everything
What is the Baron and Kenny approach to mediation?
Independent regressions of the IV to DV, IV to MV, and IV + MV to DV. Mediation occurs if the effect of the IV is reduced when MV is introduced, but only if the IV was significant in the first place
What is the basic idea behind Principal Component Analysis
○ We have multivariate data (we’ve measured participants on multiple variables)
○ We fit straight lines to our data and call the lines ‘Principal Components’ (PCs)
○ 1st PC is the best line we can fit
○ 2nd PC is second best line we can fit etc
○ Maximum number of PCs = number of variables in our dataset
○ We want to represent our data with fewer PCs
○ Correlated continuous variables, and reducing them into the least amount of factors while keeping the data
○ Aims to fit straight lines to data points
§ Second best line is line fitting the errors of the components
§ Eg reducing alcohol dryness and content into one component, while still describing the alcohol fully
□ First line reduces the diagonal distances from the data points and the principal component line
□ Worst will always be perpendicular/orthogonal to the principal component
* Can also be thought of in terms of dimensions
○ If you have n variables, you are in n dimensional space
○ Maybe there are new axes that make life simpler
○ Maybe you don’t need the full n components to describe your data well
Explain the MAP test for determine how many components to extract for PCA
○ Velicer devised a test based on partial correlations known as the Minimum Average Partial Correlation test
§ After each component Is extracted, it (and those extracted before) gets partialled out of the correlation matrix of original variables, and the average of the resulting partial correlations are calculated
§ As more correlations are partialled out, the partial correlations approaches 0
§ But at some point components that reflected ‘noise’ would be partialled out so the average partial correlation would begin to rise
§ Choose number of components corresponding to minimum average partial correlation
Describe component rotation
○ With n variables, you can have n components
§ These are completely determined and follow directly from the matrix operations
○ But if we only use smaller number of components, there is some freedom in the final solution
§ In particular, we can rotate components to get a simple structure
§ Axes get rotated until the variables are tending to load on one component only, to as great an extent that is possible, and they are as close to 0 on the other components as possible
§ With large component loadings or some variables, and small component loadings for others
§ Used more in FA than PCA