Factor Analysis Flashcards

1
Q

Factor analysis

A

Ways of chunking data into theoretically reverent sub-scales or factors

SPSS lists it as ‘data reduction method’

Way of summarising large amount of info in terms of its underlying dimensions or factors

Like a quantitative thematic analysis- runs correlations between all items in large data set

Then generates themes (components/factors) based on the patterns (which items correlation more/less)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

When would you use factor analysis/principal component analysis

A

When developing new scale/questionnaire-identify sub-scales & superfluous/irrelevant items

When trying to reduce number of items on existing questionnaire

When testing underlying structure of dataset

To see if there might be common theoretical underpinning or set of separate measures

Individual differences literature

Not when you want to chunk groups of pp together- cluster analysis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

(Exploratory) factor analysis vs PCA

A

SPSS treats them like they’re the same

Most often give same results so terms used interchangeably

Confirmatory factor analysis is different & SPSS can’t do it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Stages of exploratory factor analysis

A

1) Make correlation matrix (of all variables)
2) extract principal components
3) rotate principal components
4) interpret components/factors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Is it quantitative?

A

Yes in as far as calculations, rotations & sorting factors- based on numerical data & statistical transformations

But no critical values so no significant/non-significant or effect sizes

Factors named & interpreted far more like qualitative analysis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Types of factor analysis

A

Exploratory= looks for emergent structure in current dataset based on patterns of correlations between variables (bottom-up)

Confirmatory= tests whether factor structure in current dataset fits specified model (top-down)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Stage 1: correlation matrix

A

Straightforward correlation matrix of variables but rather than correlating mean or sum scales, this is a correlation matrix of every item in dataset

Factor analysis tries to identify underlying patterns or dimensions that explain patterns of correlations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Stage 2: extracting components

A

Extracting first component is like finding a line of best fit in regression

Trying to explain as much variance between variables as possible with single factor

This new component will be as highly correlated as possible with as many items as possible- explains most variance

Second component will try to explain as much variable as possible while controlling for the first component & so on- explains less variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Stage 3; rotation

A

After components/factors extracted, they are rotated in statistical space

Puts number of stats constraints on what final factors look like

These different rotations (patterns of constraints) may make more or less sense to use depending on data & questions asked

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Types of rotation

A

Orthogonal= factors resulting from analysis effectively cannot correlate with one another, each factor is theoretically distinct

Oblique= factors can correlate with each other, so different factors can account for some of same underlying variability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Factor loadings

A

A component/factor is way of reducing set of related variables into simpler, single score

But each of these items will have different strengths of relationships with component factor

Relationship of each item to given component/factor is called factor loading

We care about factor loadings of 0.3 or higher

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Orthogonal analyses

A

VARIMAX= maximises variance explained

QUARTIMAX= variables will load highly on one factor but low on others

EQUAMAX= seeks more equal variance explained by each factor

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Oblique analyses

A

DIRECT OBLIMIN= gives higher Eigen values but can be difficult to interpret due to overlapping constructs

PROMAX= same

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Deciding number of factors: Kaiser Criterion

A

Kaiser Criterion= Eigenvalues>1

Eigenvalues represent total (squared) factor loadings onto single factor
So higher factor loadings mean factor explains more of variance in sample

Ratio of total variance explained to number of items
So if Eigenvalues=1 for a factor, it explains same amount of variance as individual variables do by themselves

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Deciding number of factors: Scree plot

A

Look for knee or elbow of plot

Sometimes may be ambiguous

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Deciding number of factors; common sense

A

Does interpretation make theoretical sense?

e.g. is 6th factor a confusing mix of unrelated items?

Will probably use combination of methods

17
Q

Stage 4: interpreting factors

A

Once factors extracted, you use items that load onto that factor to name it

Can use the strength of factor loadings to inform this

18
Q

Requirements & assumptions

A
  • assume linear relationships between all variables
  • items/measure are all continuous variables/data
  • assumes any relationships between items are theoretically pertinent (doesn’t take into account spurious correlations)
19
Q

Limitations/requirements of factor analysis

A

Notorious for needing large sample sizes to make sense so can be hard to recruit for

Not a very reliable analysis
Very possible that you’ll get different factor structure with different samples or in same sample
Interpretation of factors can be very subjective
So normally needs multiple replications to be convincing

20
Q

Sample size

A

Common rule of thumb is N>100
Not necessarily needed but around do to start with

Kaiser-Meyer-Olkin test for sampling adequacy should be above .6 & ideally between .8-1

Communality should be high- >.6, mean commonalities should be >/=.7

21
Q

Communalities

A

By generating correlation matrix of all variables, able to calculate how much of the variance for single item is shared with variance of other items in dataset

Communality score for an item gives proportion of how much of its variance is common/shared with other items

High communality means PCA will with better as groups components based on shared variance among items so don’t need as many pp

Mean communality should be >/=.7 & all communality should be >.6