Discussion about this post

User's avatar
Andrew Cutler's avatar

Note that much of the debate in the field is about whether or not GFP even exists. For example a 2013 paper opened “The overwhelmingly dominant view of the GFP is that it represents an artefact due either to evaluative bias or responding in a socially desirable manner.” (The general factor of personality: Substance or artefact?)

If it is an artifact, that would also invalidate the Big Five. GFP can be extracted from word data (via word vectors or surveys, the results are the same). Alternately, one can extract the first five PCs, and then use varimax rotation. In this rotation, significant portions of PC1 (the GFP) are distributed to PCs 3, 4, and 5 to make Conscientiousness, Neuroticism, and Openness. If GFP is an artifact, why do that?

Even more fundamentally, if PC1 is just an artifact, it makes for a strange version of the Lexical Hypothesis: "personality structure = language structure (minus the first PC)."

Expand full comment
Norman G. Angleson's avatar

A set of personality factors should be said to be supported by factor analysis as far as it achieves simple structure rather than model fit. You can get equivalent fit for wildly different solutions [a], but the solutions which better display simple structure should be more interpretable and replicable in CFA (and I'd reckon, also more heritable).

I know varimax imposes an artificial orthogonality requirement and that many oblique rotation methods are unsatisfactory/subjective due to them relying on researcher hypotheses about how correlated a solution should be, but I don't think that it inherently has to be that oblique rotation has this flaw.

If we relax the requirement that factors be uncorrelated, a more modest criterion for simple structure might be that the correlations between factors can account for any examples of a variable correlating substantially with more than one factor.

Now, given the existence of some factor, how might we go about creating some variable for that factor which correlates with the other factors as closely as possible to what you'd predict from the variable's correlation with its 'parent' factor? The variable which does the best possible job of this is just a clone of the factor itself (or at least, a clone with some measurement error added in which doesn't correlate with any other factor or variable). As such, if we took all of the variables in a set that we think 'clumps' together, pretended that none of the other variables existed, and then took the first principal component of these variables, it's tautological to say that this first principal component is the linear combination of these variables which best facilitates simple structure.

If then, we looked at all the possible ways to group the variables and then select the grouping which minimizes the correlations between the principal components of its groups (or which minimizes the ratio of the variable-factor correlations to the factor-factor correlations; this is probably preferable but more cumbersome to calculate), it's safe to say that this would be among the best representations of the structure of our dataset, mirroring the graphical solutions that a researcher might come to, and letting the data "speak for itself".

Do this with likert self-ratings of a maximally-diverse set of actual behaviors rather than assessments of how similar people rate various adjectives to be, and I think you'd have a tough time coming up with a better model of personality.

As for testing for hierarchicalness,

A) It probably really depends that you get the rotations right

B) I'd be interested to see it looked at by an agglomerative approach rather than a divisive one. Posit next to as many oblique factors as could possibly be justified, and then when reduced to fewer factors, if it's always the case that a single factor in a given layer is responsible for all consolidation that happens relative to a lower one, then your data wasn't very hierarchical.

[a] NOTE: It's trivially the case that model fit can be improved in an exploratory context by just extracting more factors or principal components, but when people try to come up with rules about how many components to extract, this is never what they base the decision on. For a scree plot for example, it's not about which number of factors explains the most variance, it's about where (if anywhere) there's a sudden change of slope in the relationship between factor count and explained variance. With a set amount of factors to be extracted as a given however, all orthogonal rotations will be equivalent in a model fit sense even if they'd behave differently in a test for hierarchicalness or differ in replicability.

The optimal level of aggregation should be considered to be the one which performs the best when assessed for simple structure, with higher and lower levels also having validity if hierarchicalness is displayed. The only possible reason a test of model fit wouldn't conclude that more is better is if there's a punishment for complexity which concludes that at some point, model fit wasn't improved enough to justify the added complexity, and I'm dubious about how well anybody assesses complexity. Existing complexity punishments would, for example, conclude that hierarchical g models are worse than the oblique models they're derived from even when the difference in absolute fit between the two is statistically insignificant and practically negligible; if there's no difference in absolute fit but the hierarchical model is punished more for "complexity", then your measure of complexity is obviously retarded from a theoretical perspective even if it's sensible from an NHST perspective.

Expand full comment
8 more comments...

No posts