Oftentimes our questionnaire is based on an existing theoretical framework that consists of various constructs in some proposed relationships. For example, the Theory of Planned Behavior (TPB) proposes that Attitude, Subjective Norm, and Perceived Behavioral Control are the factors of Intention, which in turn affects Behavior.
The Attitude, Subjective Norm, … etc. are the constructs or factors in this model. For each construct / factor we have a number of questionnaire items that collect data about that construct / factor (in contrast to the case of linear regression where we have only one variable for each construct). For example:
[Factor 1] Attitude:
Q1: xxxx
Q2: xxxx
Q3: xxxx
[Factor 2] Subjective Norm:
Q4: xxxx
Q5: xxxx
…
This categorization usually comes from theories developed in other previous studies. But how do we know if this categorization is consistent with our data? How do we know if Q1, Q2, Q3 really measure the same factor (however it is named), and Q4, Q5,... really measure another factor, etc.?
To answer this question, we conduct a Confirmatory Factor Analysis (CFA) to confirm that the items measure their corresponding factors.
Note #1: Attitude, Subjective Norm, …, etc. are also called latent variables here because we do not measure these directly, but instead we measure them through Q1, Q2, Q3, …, etc.
Note #2: Desirably, the Q1, Q2, … should be continuous numbers. But it is a common practice to accept ordinal variables as well if we have multiple items under each construct / factor.
Note #3: Read more about CFA here: https://towardsdatascience.com/confirmatory-factor-analysis-theory-aac11af008a6.
The standard installation of SPSS cannot do CFA. You need to install AMOS, separately sold by IBM, for that purpose.
Analyses -> Factor -> Confirmatory Factor Analysis: Put the variables into their corresponding factors according to your theoretical framework.
Under Estimates, select Standardized estimate.
Under Model Fit, select the fit measures that you want. If you have no idea what to choose, just follow the default, or follow the list of measures reported in your key literatures.
Optionally, under Additional Output, select Path diagram if needed.
The Factor Loadings table shows how well the items (called indicators here) “load” to the Factors. The factor loadings are displayed in the Stand. Estimate column, while the p-value shows their statistical significance. A statistically significant factor loading >0.7 means that the item relates well to the corresponding factor. On the other hand, a small factor loading indicates that the items might not belong to that factor and should be removed from the analysis, or if it is a pilot test, the item should be revised before the main study.
The Model Fit tables show you the fit measures of the model. One should report some of these fit measures and compare them with the reference values when reporting the CFA. For example, CFI should preferably be >0.95. SRMR should preferably be <0.09, etc. Check the literature for a list of reference values to be used for comparison here.