Exploratory Factor Analyses Flashcards
When do you carry out a factor analysis?
When you have continuous latent data and continuous observed data
How did we know that the observed data was categorical in IRT?
Answeres were yes/ no, likert scale etc
How do you know if the observed data is continuous?
In psychology we rarely have truly continuous data, an example is reaction time. As a rule; if an item has more than five points (scale) and forms a normal distribution, you can consider it as a continuous item and perform factor analysis on it.
When is factor analysis often applied?
Sum scores on sub tests (e.g dimensions of intelligence)
What is the exciting thing about factor analysis as compared to item response theory according to Dylan
The nice thing about IRT is that you really analyse the individual items. FA is more flexible; continuous data is easier to model and the mathematics and formulas is more simple.
Why did IRT form an S shaped curve?
Because you’re modelling a probability of an outcome (correct score) since you have categorical data
Does one-factor FA have a s shaped curve? Explain
No, it’s a linear function. The expected value is on the x axis and since we’re no longer estimating probability and working with continuous data, the scores go higher than one and we can use a linear model. For this reason it is sometimes known as the linear factor model.
Explain the linear factor analysis equation
𝐸 (𝑋𝑝𝑖 |𝜃𝑝) =𝑏𝑖 +𝑎𝑖𝜃𝑝
or 𝑋𝑝𝑖 =𝑏𝑖 +𝑎𝑖𝜃𝑝 +𝜖𝑝𝑖
with VAR(Xpi | 𝜃p)
𝑏𝑖 is commonly referred to as item attractiveness but it roughly translates to the IRT item difficulty/ easiness parameter and it is the intercept in the slope.
e. g., “I think about suicide” has a low attractiveness
e. g., “I am statisfied with my life” has a higher attractiveness
𝑎𝑖 is the item discrimination, same as IRT and it forms the slope of the function. We model the expected value of the continuous item, and if it is a good model then the observations should be around that line.
the variance is just how much the observed points vary for the modelled line
How can the notation for this model change? Explain
𝐸 (𝑋𝑝𝑖 |𝜃𝑝) =𝑏𝑖 +𝑎𝑖𝜃𝑝 is written as: 𝑌𝑝𝑖 =𝜐𝑖 +𝜆𝑖𝜂𝑝 +𝜖𝑝𝑖 in factor analysis literature 𝜐𝑖 is an intercept 𝜆𝑖 is a factor loading 𝜂𝑝 is the common factor 𝜖𝑝𝑖 is the residual
They mean the same things but are just written differently in IRT literature compared to FA literature.
Conceptually, what is the goal of factor analysis?
Its a statistical approach to extract common variance from the items and separate it from the variable specific effects
How do the parameters translate to the variance measured?
𝑌𝑝𝑖 =𝜐𝑖 +𝜆𝑖𝜂𝑝 +𝜖𝑝𝑖
The common factor variance (𝜎^2𝜂) is the variance caused by the latent trait
The factor loadings (𝜆𝑖) tunes how much common factor variability comes from each item out of all the variability in the items - how well each item measures the latent trait
The intercepts (𝜐𝑖), in a single group application, are simply the item means
The residual variances (𝜎^2,𝜖𝑖) tells you how much of the variance is unique to the item
What does this model imply about the data? How can this be used in calculations?
The model implies some kind of structure in the data in terms of the variance. You can calculate how much variability does this model predict for item 1. You can compare this to the observed variance which should be the factor loading squared times the variance plus the residual variance.
It also implies some kind of covariance between the items since they measure the same thing. If you score high on one item you are assumed to score higher on a second item. You can calculate the expected covariance through this.
How do you calculate the expected covariance? What should you look for in this?
The expected covariance is the first factor loading multiplied by the second factor loading etc multiplied by the factor variance. You should look to see if this is close to the observed covariance to assess whether this is a good model for the data.
How is the proportion of the variance calculated?
𝜌|2𝑌|𝑖𝜂𝑝 = 𝜆|2𝑖| 𝜎|2𝜂| / 𝜆|2𝑖| 𝜎|2𝜂| +𝜎|2𝜖|𝑖
i.e the variance of the latent variable multipled / the total variance = the variance explained by the factor
The varaince not explained by the factor (uniqueness) is calculated by:
1 - 𝜌|2𝑌|𝑖𝜂𝑝 = 𝜎|2 𝜖𝑖| / 𝜆|2𝑖| 𝜎|2𝜂| +𝜎|2𝜖|𝑖
However most of the time you can just read these from the output in Rstudio
What is the equivalent of chronbachs alpha from CTT here and hoiw is it calculated?
The reliability of the sum score, the ratio between the variability due to the latent trait and the total variability
𝜌𝑌 =𝜎^2 (𝑇 ) / 𝜎^2 (𝑌) = 𝜎^2 (𝐸(𝑌)) / 𝜎^2 (𝑌 )
= …
= 𝜎𝜂2 × (E|𝑛,𝑖=1| 𝜆𝑖)^ 2 /
𝜎|2,𝜂| × (E|𝑛,𝑖=1| 𝜆𝑖)^ 2 + E|𝑛,𝑖=1 𝜎|2,𝜖𝑖|
i.e multiply the factor variance by the sum of the loadings squared, divided by the factor variance by the loadings squared by the sum of the residual variances squared summed
How does identification change compared to IRT and why?
It is very similar to IRT, we need to identify the model because the latent variable doesn’t have a scale or a unit so we have to create one. In factor analysis, however, we play around with this more. In IRT we fixed the mean to 0 and the std to 1 because the R packages don’t allow you to change it much and its not interesting when its one dimensional. People like to change the identification to get a different scale for the parameters, this will not change the conclusions or the p-value since the proportions between the parameters don’t change.
Give an example of how you can have two different options for identification in factor analysis
Option 1:
• 𝜇𝜂 =0
• 𝜎𝜂2 =1
Option 2:
• 𝜇𝜂 =0
• Fix one factor loading to 1
By picking an arbitrary factor loading and fixing it to one is like saying that the scale of the latent variable is the same as the scale of that item
What is the most dominant approach to parameter estimation in factor analysis?
Maximum likelihood
What does using MLE in this instance assume about the data?
Normally distributed
What can you do as opposed to fitting the model on the raw data with MLE for a factor analysis? Why might you want to do this?
You have the option to only analyse the observed covariance matrix which is very useful for factor analysis since a covariance matrix already contains all the information about the structure of your data. From a covariance matrix you ca already fit a one parameter model since you have your factor loadings and residual variance.
What is a downside to using the covariance matrix in estimating the MLE for a factor analysis?
There’s no intercepts in the model because for the intercepts you really need the means of the data which is not contained in a covariance matrix.
What two alternative methods for parameter estimation exist?
Weighted least squares and bayesian estimation (Also popular but not discussed in this course)
Given data with 1000 subjects answering 10 questions on a 5 point likert scale, how would you write up code to carry out a factor analysis? Explain the code
head(E)
library(lavaan)
model = “eta = ~Y1 + Y2 + Y3 + Y4 +Y5 + Y6 +Y7 + Y8 +Y9 + Y10”
fit = cfa(model = model, data = E, meansstructure = TRUE, std.lv = TRUE)
where eta is the common factor and can be given any name akin to a variable, =~ indicated that it is measured by…, cfa runs a confirmatory factor analysis, data is called E, meanstructure = TRUE means that you want to calculate the intercepts, std.lv = TRUE means that you want to standardise the latent variable with 𝜇𝜂 =0 and 𝜎𝜂2 =1