Processing math: 100%
+ - 0:00:00
Notes for current slide
Notes for next slide

WEEK 4
Exploratory Factor Analysis 1

Data Analysis for Psychology in R 3

dapR3 Team

Department of Psychology
The University of Edinburgh

1 / 47

Learning Objectives

  1. Understand the conceptual difference between PCA and EFA
  2. Understand the strengths and weaknesses of different methods to estimate the optimal number of factors
  3. Understand the purpose of factor rotation and the difference between orthogonal and oblique rotation.
  4. Run and interpret EFA analyses in R.
2 / 47

Part 1: Introduction to EFA

Part 2: EFA vs PCA

Part 3: Estimation & Number of factors problem

Part 4: Factor rotation

Part 5: Example and interpretation

3 / 47

Real friends don't let friends do PCA.

W. Revelle, 25 October 2020

4 / 47

Questions to ask before you start

  • Why are your variables correlated?

    • Agnostic/don't care
    • Believe there are underlying "causes" of these correlations
  • What are your goals?

    • Just reduce the number of variables
    • Reduce your variables and learn about/model their underlying (latent) causes

5 / 47

Questions to ask before you start

  • Why are your variables correlated?

    • Agnostic/don't care
    • Believe there are underlying "causes" of these correlations
  • What are your goals?

    • Just reduce the number of variables**
    • Reduce your variables and learn about/model their underlying(latent) causes

6 / 47

Latent variables

  • One of many features that distinguish factor analysis and principal components analysis

  • Key concept of psychometrics (factor analysis is a part)

  • Theorized common cause (e.g., cognitive ability) of responses to a set of variables

    • Explain correlations between measured variables
    • Held to be true
    • No direct test of this theory

7 / 47

An example

  • Suppose we conduct a survey, and we ask 1000 people to respond to the following 10 questions which measure different aspects aggression.
  1. I hit someone
  2. I kicked someone
  3. I shoved someone
  4. I battered someone
  5. I physically hurt someone on purpose
  6. I deliberately insulted someone
  7. I swore at someone
  8. I threatened to hurt someone
  9. I called someone a nasty name to their face
  10. I shouted mean things at someone
  • Read the items, how do you think they might group?
8 / 47

Our running example

  • What do you see if you look at the item correlations?
round(cor(agg.items),2)
## item1 item2 item3 item4 item5 item6 item7 item8 item9 item10
## item1 1.00 0.51 0.48 0.39 0.55 0.00 0.09 0.05 0.09 0.05
## item2 0.51 1.00 0.55 0.45 0.61 0.04 0.11 0.08 0.09 0.04
## item3 0.48 0.55 1.00 0.44 0.58 0.04 0.11 0.07 0.09 0.04
## item4 0.39 0.45 0.44 1.00 0.48 0.03 0.11 0.04 0.10 0.01
## item5 0.55 0.61 0.58 0.48 1.00 0.01 0.09 0.02 0.09 0.01
## item6 0.00 0.04 0.04 0.03 0.01 1.00 0.52 0.53 0.42 0.43
## item7 0.09 0.11 0.11 0.11 0.09 0.52 1.00 0.74 0.56 0.57
## item8 0.05 0.08 0.07 0.04 0.02 0.53 0.74 1.00 0.54 0.57
## item9 0.09 0.09 0.09 0.10 0.09 0.42 0.56 0.54 1.00 0.42
## item10 0.05 0.04 0.04 0.01 0.01 0.43 0.57 0.57 0.42 1.00
9 / 47

Practical Steps

  • So how do we move from data and correlations to a factor analysis?
    • We will be working through these steps in the next 2 weeks.
  1. Check the appropriateness of the data and decide of the appropriate estimator.
  2. Decide which methods to use to select a number of factors.
  3. Decide conceptually whether to apply rotation and how to do so.
  4. Decide on the criteria to assess and modify your solution.
  5. Run the analysis.
  6. Evaluate the solution (apply 4)
  7. Select a final solution and interpret the model, labeling the factors.
  8. Report your results.
10 / 47

What does an EFA look like?

library(psych)
agg_res <- fa(agg.items, nfactors = 2, fm = "ml", rotate = "oblimin")
agg_res
## Factor Analysis using method = ml
## Call: fa(r = agg.items, nfactors = 2, rotate = "oblimin", fm = "ml")
## Standardized loadings (pattern matrix) based upon correlation matrix
## ML1 ML2 h2 u2 com
## item1 0.00 0.67 0.45 0.55 1
## item2 0.02 0.75 0.57 0.43 1
## item3 0.02 0.72 0.51 0.49 1
## item4 0.01 0.60 0.36 0.64 1
## item5 -0.03 0.81 0.66 0.34 1
## item6 0.63 -0.04 0.39 0.61 1
## item7 0.85 0.04 0.74 0.26 1
## item8 0.86 -0.03 0.73 0.27 1
## item9 0.63 0.05 0.41 0.59 1
## item10 0.67 -0.04 0.44 0.56 1
##
## ML1 ML2
## SS loadings 2.71 2.56
## Proportion Var 0.27 0.26
## Cumulative Var 0.27 0.53
## Proportion Explained 0.51 0.49
## Cumulative Proportion 0.51 1.00
##
## With factor correlations of
## ML1 ML2
## ML1 1.00 0.11
## ML2 0.11 1.00
##
## Mean item complexity = 1
## Test of the hypothesis that 2 factors are sufficient.
##
## df null model = 45 with the objective function = 3.91 with Chi Square = 3894
## df of the model are 26 and the objective function was 0.02
##
## The root mean square of the residuals (RMSR) is 0.01
## The df corrected root mean square of the residuals is 0.01
##
## The harmonic n.obs is 1000 with the empirical chi square 8.33 with prob < 1
## The total n.obs was 1000 with Likelihood Chi Square = 16.15 with prob < 0.93
##
## Tucker Lewis Index of factoring reliability = 1.004
## RMSEA index = 0 and the 90 % confidence intervals are 0 0.007
## BIC = -163.4
## Fit based upon off diagonal values = 1
## Measures of factor score adequacy
## ML1 ML2
## Correlation of (regression) scores with factors 0.94 0.92
## Multiple R square of scores with factors 0.88 0.85
## Minimum correlation of possible factor scores 0.77 0.70
11 / 47

Key parts of the output

  • Factor loading's, like PCA loading's, show the relationship of each measured variable to each factor.

    • They range between -1.00 and 1.00
    • Larger absolute values = stronger relationship between measured variable and factor
  • We interpret our factor models by the pattern and size of these loading's.

    • Primary loading's: refer to the factor on which a measured variable has it's highest loading
    • Cross-loading's: refer to all other factor loading's for a given measured variable
  • Square of the factor loading's tells us how much item variance is explained ( h2 ), and how much isn't ( u2)

  • Factor correlations : When estimated, tell us how closely factors relate (see rotation)

  • SS Loading and proportion of variance information is interpreted as we discussed for PCA.

12 / 47

Part 1: Introduction to EFA

Part 2: EFA vs PCA

Part 3: Estimation & Number of factors problem

Part 4: Factor rotation

Part 5: Example and interpretation

13 / 47

PCA versus EFA: How are they different?

  • PCA

    • The observed measures (x1,x2,x3) are independent variables
    • The component (z) is the dependent variable
    • Explains as much variance in the measures (x1,x2,x3) as possible
    • Components are determinate
  • EFA

    • The observed measures (y1,y2,y3) are dependent variables
    • The factor (ξ), is the independent variable
    • Models the relationships between variables (ry1,y2,ry1,y3,ry2,y3)
    • Factors are indeterminate
14 / 47

Modeling the data

  • What does it mean to model the data?

  • EFA tries to explain these patterns of correlations

    • If, for a set of items (here 3), the model (factor or ξ) is good, it will explain their interrelationships
    • Read the dots () as "given" or "controlling for"

ρ(y1,y2ξ)=corr(e1,e2)=0ρ(y1,y3ξ)=corr(e1,e3)=0ρ(y2,y3ξ)=corr(e2,e3)=0

15 / 47

Modeling the data

  • In order to model these correlations, EFA looks to distinguish between the true and unique item variance

var(total)=var(common)+var(specific)+var(error)

  • True variance
    • Variance common to an item and at least one other item
    • Variance specific to an item that is not shared with any other items
  • Unique variance
    • Variance specific to an item that is not shared with any other items
    • Error variance
16 / 47

The general factor model equation

Σ=ΛΦΛ+Ψ

Σ: A p×p observed covariance matrix (from data)

Λ: A p×m matrix of factor loading's (relates the m factors to the p items)

Φ: An m×m matrix of correlations between factors ("goes away" with orthogonal factors)

Ψ: A diagonal matrix with p elements indicating unique (error) variance for each item

17 / 47

Assumptions

  • As EFA is a model, just like linear models and other statistical tools, it has some assumptions:

    1. The residuals/error terms (e) should be uncorrelated (it's a diagonal matrix, remember!)
    2. The residuals/errors should not correlate with factor
    3. Relationships between items and factors should be linear, although there are models that can account for nonlinear relationships
18 / 47

End of Part 2

19 / 47

Part 1: Introduction to EFA

Part 2: EFA vs PCA

Part 3: Suitability of data, Estimation & Number of factors problem

Part 4: Factor rotation

Part 5: Example and interpretation

20 / 47

Data suitability

  • This boils down to is the data correlated.

    • So the initial check is to look to see if there are moderate correlations (roughly > .20)
  • We can take this a step further and calculate the squared multiple correlation (SMC).

    • SMC are multiple correlations of each item regressed on all p1 other variables
    • this metric tells us how much shared variation there is between an item and all other items
    • This is one way to estimate a communalities (see later)
  • There are also some statistical test (e.g. Bartlett's test)

    • However, these tests are generally not that informative.
21 / 47

Estimation

  • For PCA, we discussed the use of the eigen-decomposition.

    • This is not an estimation method, it is simply a calculation
  • As we have a model for the data in factor analysis, we need to estimate the model parameters

    • primarily here the factor loading's.
22 / 47

Estimation & Communalities

  • The most efficient way to factor analyze data is to start by estimating communalities

    • Communalities are estimates of how much true variance any variable has
    • Indicate how much variance in an item is explained by other variables, or factors
  • If we consider that EFA is trying to explain true common variance, then communalitie estimates are more useful to us than total variance.

  • Estimating communalities is difficult because population communalities are unknown

    • Range from 0 (no shared variance) to 1 (all variance is shared)
    • Occasionally estimates will be 1 (called a 'Heywood Case')
    • Methods often are iterative and "mechanical" as a result
23 / 47

Principal axis factoring

  • This approach to EFA uses squared multiple correlation (SMC)
  1. Compute initial communalities from SMCs

  2. Once we have these reasonable lower bounds, we substitute the 1s in the diagonal of our correlation matrix with the SMCs derived in step 1

  3. Obtain the factor loading matrix using the eigenvalues and eigenvectors of the matrix obtained in the step 2

  • Some versions of principal axis factor use an iterative approach in which they replace the diagonal with the communalities obtained in step 3, and then repeat step 3, and so on, a set number of times
24 / 47

Method of minimum residuals

  • This is an iterative approach and the default of the fa procedure
  1. Starts with some other solution, e.g., PCA or principal axes, extracting a set number of factors

  2. Adjusts loading's of all factors on each variable so as to minimize the residual correlations for that variable

  • MINRES doesn't "try" to estimate communalities

  • If you apply principal axis factoring to the original correlation matrix with a diagonal of communalities derived from step 2, you get the same factors as in the method of minimum residuals

25 / 47

Maximum likelihood estimation

  • Uses a general iterative procedure for estimating parameters that we have previously discussed.

    • The procedure works to find values for these parameters that maximize the likelihood of obtaining the covariance matrix
  • Method offers the advantage of providing numerous "fit" statistics that you can use to evaluate how good your model is compared to alternative models

    • Recall we can compare the model implied to the actual covariances to get fit.
  • Assume a distribution for your data, e.g., a normal distribution

26 / 47

ML con's

  • The issue is that for big analyses, sometimes it is not possible to find values for factor loadings that = MLE estimates.

    • Referred to as non-convergence (you may see warnings)
  • Also may produce solutions with impossible values

    • Factor loadings > 1.00 (Heywood cases), thus negative residuals.
    • Factor correlations > 1.00
27 / 47

Non-continuous data

  • Sometimes the construct we are interested in is not continuous, e.g. number of crimes committed.

  • Sometimes we assume the construct is, but we measure it with a discrete scale.

  • Most constructs we seek to measure by questionnaire fall into the latter category.

    • It's thus usually okay to treat the data as if they are, too
    • The exception is for maximum likelihood factor analysis
    • Or if the observed distribution is very skewed
28 / 47

Non-continuous data

29 / 47

Non-continuous data

  • If we are concerned and the construct is normally distributed, we can conduct our analysis on a matrix of polychoric correlations

  • If the construct is not normally distributed, you can conduct a factor analysis that allows for these kinds of variables

30 / 47

Choosing an estimator

  • The best option, as with many statistical models, is ML.

  • If ML solutions fail to converge, principal axis is a simple approach which typically yields reliable results.

  • If concerns over the distribution of variables, use PAF on the polychoric correlations.

31 / 47

Number of factors

  • We have discussed the methods for deciding on the number of factors in the context of PCA.

  • Recall we have 4 tools:

    • Variance explained
    • Scree plots
    • MAP
    • Parallel Analysis (FA or PCA)
  • For FA, we generally want a slightly more nuanced approach than pure variance:

    • Use them all to provide a range of plausible number of factors
    • Treat MAP as a minimum
    • PA as a maximum
    • Explore all solutions in this range and select the one that yields the best numerically and theoretically.
  • We will go through this process in later video.

32 / 47

End of Part 3

33 / 47

Part 1: Introduction to EFA

Part 2: EFA vs PCA

Part 3: Estimation & Number of factors problem

Part 4: Factor rotation

Part 5: Example and interpretation

34 / 47

Factor rotation: what and why?

  • Factor solutions can sometimes be complex to interpret.

    • the pattern of the factor loading's is not clear.
    • The difference between the primary and cross-loading's is small
  • Why is this the case?

    • Rotational indeterminacy means that there are an infinite number of pairs of factor loading's and factor score matrices which will fit the data equally well, and are thus indistinguishable by any numeric criteria
  • In other words, there is no unique solution to the factor problem

  • And this is also in part why the theoretical coherence of the models plays a much bigger role in FA than PCA.

35 / 47

Analytic rotation

  • Factor rotation is an approach to clarifying the relationships between items and factors.

    • Rotation aims to maximize the relationship of a measured item with a factor.
    • That is, make the primary loading big and cross-loading's small.
  • Thus although we can not numerically tell rotated solutions apart, we can select the one with the most coherent solution.

  • There are many different ways this can be achieved.

    • One framework for this is referred to as simple structure
36 / 47

Simple structure

  • Adapted from Sass and Schmitt (2011):
  1. Each variable (row) should have at least one zero loading

  2. Each factor (column) should have same number of zero’s as there are factors

  3. Every pair of factors (columns) should have several variables which load on one factor, but not the other

  4. Whenever more than four factors are extracted, each pair of factors (columns) should have a large proportion of variables which do not load on either factor

  5. Every pair of factors should have few variables which load on both factors

37 / 47

Orthogonal vs Oblique Rotation

  • All factor rotation methods seek to optimize one or more aspects of simple structure.

    • But there are two broad groupings
  • Orthogonal

    • Includes varimax and quartimax rotations
    • Axes at right angles; correlations between factors are zero
  • Oblique

    • Includes promax and oblimin rotations
    • Axes are not at right angles; correlations between factors are not zero
38 / 47

The impact of rotation

Original correlations

EFA with no rotation and 5 factors

39 / 47

The impact of rotation

EFA with no rotation and 5 factors

EFA with orthogonal rotation and 5 factors

40 / 47

The impact of rotation

EFA with orthogonal rotation and 5 factors

EFA with oblique rotation and 5 factors

41 / 47

How do I choose which rotation?

  • Easy, my recommendation is always to choose oblique.

  • Why?

    • It is very unlikely factors have correlations of 0
    • If they are close to zero, this is allowed within oblique rotation
    • The whole approach is exploratory, and the constraint is unnecessary.
  • However, there is a catch...

42 / 47

Interpretation and obique rotation

  • When we have an obliquely rotated solution, we need to draw a distinction between the pattern and structure matrix.

    • Pattern Matrix: matrix of regression weights (loading's) from factors to variables.
    • Structure Matrix: matrix of correlations between factors and variables.
  • When we use orthogonal rotation, the pattern and structure matrix are the same.

  • When we use oblique rotation, the structure matrix is the pattern matrix multiplied by the factor correlations.

  • In most practical situations, this does not impact what we do, but it is important to highlight the distinction.

43 / 47

End of Part 4

44 / 47

Part 1: Introduction to EFA

Part 2: EFA vs PCA

Part 3: Estimation & Number of factors problem

Part 4: Factor rotation

Part 5: Example and interpretation

45 / 47

Worked Example

  • In this weeks LEARN folder there is a worked example of an EFA.
46 / 47

End

47 / 47

Learning Objectives

  1. Understand the conceptual difference between PCA and EFA
  2. Understand the strengths and weaknesses of different methods to estimate the optimal number of factors
  3. Understand the purpose of factor rotation and the difference between orthogonal and oblique rotation.
  4. Run and interpret EFA analyses in R.
2 / 47
Paused

Help

Keyboard shortcuts

, , Pg Up, k Go to previous slide
, , Pg Dn, Space, j Go to next slide
Home Go to first slide
End Go to last slide
Number + Return Go to specific slide
b / m / f Toggle blackout / mirrored / fullscreen mode
c Clone slideshow
p Toggle presenter mode
t Restart the presentation timer
?, h Toggle this help
oTile View: Overview of Slides
Esc Back to slideshow