Parallel analysis
Statistical method
From Wikipedia, the free encyclopedia
Parallel analysis, also known as Horn's parallel analysis, is a statistical method used to determine the number of components to keep in a principal component analysis or factors to keep in an exploratory factor analysis. It is named after psychologist John L. Horn, who created the method, publishing it in the journal Psychometrika in 1965.[1] The method compares the eigenvalues generated from the data matrix to the eigenvalues generated from a Monte-Carlo simulated matrix created from random data of the same size.[2] Horn developed the method as a correction to the Kaiser criterion, arguing that sampling error and least-squares capitalization inflate eigenvalues in sample data, causing the Kaiser rule to overestimate the number of factors to retain.[1]
Evaluation and comparison with alternatives
Parallel analysis is regarded as one of the more accurate methods for determining the number of factors or components to retain. In particular, unlike early approaches to dimensionality estimation (such as examining scree plots), parallel analysis has the virtue of an objective decision criterion.[3] In a Monte Carlo simulation study across 96 conditions with known component structure, Zwick and Velicer found that parallel analysis and Velicer's MAP test[4] generally outperformed the other methods tested, while the Kaiser rule tended to severely overestimate the number of components.[3]
Since its original publication, multiple variations of parallel analysis have been proposed.[5][6] Other methods of determining the number of factors or components to retain in an analysis include the scree plot, Kaiser rule, or Velicer's MAP test.[4]
Anton Formann provided both theoretical and empirical evidence that parallel analysis's application might not be appropriate in many cases since its performance is influenced by sample size, item discrimination, and type of correlation coefficient.[7]
An extensive 2022 simulation study by Haslbeck and van Bork[8] found that parallel analysis was among the best-performing existing methods, but was slightly outperformed by their proposed prediction error-based approach.
Implementation
Parallel analysis has been implemented in JASP, SPSS, SAS, STATA, and MATLAB[9][10][11] and in multiple packages for the R programming language, including the psych[12][13] multicon,[14] hornpa,[15] and paran packages.[16][17] Parallel analysis can also be conducted in Mplus version 8.0 and forward.[18]