Hostname: page-component-6bf8c574d5-7jkgd Total loading time: 0 Render date: 2025-02-20T23:16:51.529Z Has data issue: false hasContentIssue false

How Data Analysis Can Dominate Interpretations of Dominant General Factors

Published online by Cambridge University Press:  02 October 2015

Brenton M. Wiernik*
Affiliation:
Department of Psychology, University of Minnesota
Michael P. Wilmot
Affiliation:
Department of Psychology, University of Minnesota
Jack W. Kostal
Affiliation:
Department of Psychology, University of Minnesota
*
Correspondence concerning this article should be addressed to Brenton M. Wiernik, Department of Psychology, University of Minnesota, Minneapolis, MN 55455. E-mail: wiernik@workpsy.ch
Rights & Permissions [Opens in a new window]

Extract

A dominant general factor (DGF) is present when a single factor accounts for the majority of reliable variance across a set of measures (Ree, Carretta, & Teachout, 2015). In the presence of a DGF, dimension scores necessarily reflect a blend of both general and specific factors. For some constructs, specific factors contain little unique reliable variance after controlling for the general factor (Reise, 2012), whereas for others, specific factors contribute a more substantial proportion of variance (e.g., Kinicki, McKee-Ryan, Schriesheim, & Carson, 2002). We agree with Ree et al. that the presence of a DGF has implications for interpreting scores. However, we argue that the conflation of general and specific factor variances has the strongest implications for understanding how constructs relate to external variables. When dimension scales contain substantial general and specific factor variance, traditional methods of data analysis will produce ambiguous or even misleading results. In this commentary, we show how several common data analytic methods, when used with data sets containing a DGF, will substantively alter conclusions.

Type
Commentaries
Copyright
Copyright © Society for Industrial and Organizational Psychology 2015 

A dominant general factor (DGF) is present when a single factor accounts for the majority of reliable variance across a set of measures (Ree, Carretta, & Teachout, Reference Ree, Carretta and Teachout2015). In the presence of a DGF, dimension scores necessarily reflect a blend of both general and specific factors. For some constructs, specific factors contain little unique reliable variance after controlling for the general factor (Reise, Reference Reise2012), whereas for others, specific factors contribute a more substantial proportion of variance (e.g., Kinicki, McKee-Ryan, Schriesheim, & Carson, Reference Kinicki, McKee-Ryan, Schriesheim and Carson2002). We agree with Ree et al. that the presence of a DGF has implications for interpreting scores. However, we argue that the conflation of general and specific factor variances has the strongest implications for understanding how constructs relate to external variables. When dimension scales contain substantial general and specific factor variance, traditional methods of data analysis will produce ambiguous or even misleading results. In this commentary, we show how several common data analytic methods, when used with data sets containing a DGF, will substantively alter conclusions.

Job satisfaction is a quintessential multidimensional construct with a DGF. It comprises several dimensions, each of which reflects attitudes toward different components of the job and all of which are simultaneously influenced by a DGF. Thus, scores on satisfaction dimension scales reflect both general and specific attitudes, although the relative proportion of each differs across scales. Throughout this commentary, we refer to analyses conducted with job satisfaction and job performance data reported by Edwards, Bell, Arthur, and Decuir (Reference Edwards, Bell, Arthur and Decuir2008). Job satisfaction was measured using a single item measuring overall satisfaction and the Job Descriptive Index, which contains five dimension scales (Work, Pay, Promotion, Supervision, and Coworkers). Performance was measured using supervisor ratings of task performance and contextual performance. More details on the sample are available in the original article. All structural equations models (SEMs) were estimated using OpenMx version 2.0.1 (Boker et al., Reference Boker, Neale, Maes, Wilde, Spiegel and Brick2015). Results of all analyses are shown in Table 1.

Table 1. Results From Four Common Data Analytic Methods for Job Satisfaction

Note. Values for multiple regressions are observed standardized regression coefficients from the dimension scales to the criteria; values for structural equations models (SEMs) are standardized path coefficients from the latent factors to the observed criteria; values in brackets are 95% confidence intervals (bootstrapped confidence intervals for SEMs).

aZero-order correlations for the single-item overall satisfaction measure.

Common Methods for Analyzing Relations and Their Susceptibility to Misinterpretation

1. Zero-Order Correlations

The most straightforward analytic method is to examine the correlation between the external variable and each of the individual dimensions. In the presence of a DGF, this approach is problematic because all of the observed correlations reflect a composite of general and specific factor variance. Thus, a large correlation could mean that (a) the general factor influences the criterion, (b) the specific factor does, or (c) both do. Conversely, a correlation of zero could reflect that (d) neither the general nor the specific factors are related to the criterion or (e) both are, but in opposite directions. In short, zero-order correlations cannot separate general and specific factor influences, rendering interpretation ambiguous at best. Nevertheless, researchers frequently interpret dimension scale correlations as though they reflect only specific factor variance (e.g., Kinicki et al., Reference Kinicki, McKee-Ryan, Schriesheim and Carson2002).

A related issue occurs when using a composite to index the general factor. Such a sum score reflects not only the general factor but also specific factors and measurement error. As estimates of general factor influence, composite score correlations will be inflated if the general and specific factor(s) predict in the same direction (i.e., enhancing conflation) or attenuated if they predict in opposite directions (i.e., suppressive conflation). Composite correlations reflect an average criterion relation across dimensions, not simply the effect of the general factor.

On the basis of zero-order correlations from Edwards et al. (Reference Edwards, Bell, Arthur and Decuir2008), researchers would conclude that overall satisfaction and satisfaction with work itself have weak positive relations to task performance but that other scales show negligible relations (see Table 1). Similarly, one would conclude that contextual performance is weakly to moderately positively related to overall, work, pay, and supervisor satisfaction but not satisfaction with promotions or coworkers. Finally, on the basis of correlations with a satisfaction composite, researchers would conclude that overall satisfaction is positively related to contextual performance but not task performance. As the analyses below show, most of these conclusions would be wrong.

2. Multiple Regression

Multiple regression examines the combined predictive effects of dimension scores for a criterion of interest. Analyses may be performed either with observed scale scores or with latent variables in an SEM framework (but without specifying a general factor). In either case, this approach is problematic in the presence of a DGF because the influence of the general factor will create multicollinearity problems, making the pattern of regression or structural path coefficients unstable. Further, regression coefficients will primarily reflect each dimension scale's loading on the DGF, not their unique influence on the criterion. In the case of suppressive conflation, R 2 for the model will also be an underestimate, as positive and negative effects of the general and specific factors cancel out within the individual scales. Although multiple regression is affected by intercorrelations among predictors, researchers often focus their interpretations on specific factors rather than attending to the influence of the general factor (e.g., Edwards et al., Reference Edwards, Bell, Arthur and Decuir2008). From the multiple regression results in Table 1, researchers would come to similar conclusions as earlier: Overall satisfaction and work satisfaction are positively related to task performance, whereas overall satisfaction and supervisor satisfaction are positively related to contextual performance. In addition, satisfaction with promotions and coworkers shows small negative effects, suggesting that once other facets are controlled, these dimensions are negatively related to the criteria. R 2 values show that as a set, satisfaction measures explain only a moderate amount of variance in task and contextual performance.

3. General Factor SEM

A third method for assessing relations to external criteria focuses entirely on the DGF. That is, general factor SEMs attribute all predictive power to the general factor; specific factor variance is ignored entirely. Although ostensibly relevant in the presence of a DGF, this approach is essentially the same as using a sum score composite and shares many of its disadvantages. In the case of enhancing conflation, dimension loadings on the general factor and the structural path from the general factor to the criterion will be inflated. In the case of suppressive conflation, the same process will attenuate the general factor's structural coefficient. In both cases, specific effects are forced through the general factor to fit the model. Consequences of failing to specify specific factor structural paths are most severe when general factor saturation is weak. Because a general factor SEM is the latent analogue to an observed score composite, it is unsurprising that their results are virtually identical (see Table 1).

4. Bifactor Model

The preceding analytic approaches share a common limitation: General and specific factor variances are not disentangled in the predictive model, biasing the conclusions drawn about the DGF and specific factors. Bifactor modeling offers a solution (Reise, Reference Reise2012). In a bifactor model, each indicator loads on the general factor (which influences all measures) and a specific factor (which influences some measures). General and specific factors are constrained to be uncorrelated, which allows the unique predictive power of each to be examined separately. A bifactor model of Edwards et al.'s (Reference Edwards, Bell, Arthur and Decuir2008) data is shown in Figure 1.

Figure 1. Standardized loadings for a bifactor model of the Edwards et al. (Reference Edwards, Bell, Arthur and Decuir2008) job satisfaction measures with 95% bootstrapped confidence intervals.

Bifactor predictive models are best estimated via SEM and can be fit using a variety of software packages (for an alternative approach using multiple regression and residualized factor scores, see Salgado, Moscoso, & Berges, Reference Salgado, Moscoso and Berges2013). Before examining how a bifactor model affects interpretations of DGFs, a few words about how bifactor modeling works are in order. In fitting a bifactor SEM, ideally, multiple indicators for each specific factor are used (e.g., multiple items from each scale of a personality measure; McAbee, Oswald, & Connelly, Reference McAbee, Oswald and Connelly2014). Using multiple indicators allows specific factors to represent shared reliable variance, rather than a mix of reliable and error variance, and permits their simultaneous inclusion in the predictive model (cf. McAbee et al., Reference McAbee, Oswald and Connelly2014). However, if single indicators are used for specific factors, such as when only scale scores are available or when reanalyzing a published correlation matrix, at least one specific factor must be excluded from the predictive model to avoid exact linear dependence (Chen, Hayes, Carver, Laurenceau, & Zhang, Reference Chen, Hayes, Carver, Laurenceau and Zhang2012). Thus, when predicting performance using a bifactor model of Edwards et al.'s (Reference Edwards, Bell, Arthur and Decuir2008) data, we excluded the uniqueness of the single “overall satisfaction” item, based on the assumption that most of this variance was error (Wanous, Reichers, & Hudy, Reference Wanous, Reichers and Hudy1997).

Results of the bifactor analyses produce a striking pattern (see Table 1). After removing the general factor variance, all specific satisfaction factors are negatively correlated with both criteria, although they vary widely in their magnitude and precision. What this means is that the positive relation to performance comes from the DGF, overall satisfaction, not from evaluations of particular work features. For purposes of contrast, such a conclusion is precisely the opposite of that of Kinicki et al. (Reference Kinicki, McKee-Ryan, Schriesheim and Carson2002). Further, overall, promotions and coworkers satisfaction factors show far stronger relations with performance than were observed in any of the analyses wherein their variances were conflated, illustrating the relevant, but hidden, effects of suppressive conflation. Such a pattern of results is similar to bifactor analyses of other constructs (cf. Chen et al., Reference Chen, Hayes, Carver, Laurenceau and Zhang2012).

Despite their interpretive advantages, bifactor analyses can present several challenges. Two of these are worth mentioning. First, in some cases, one or more specific factors may have negligible or negative estimated variances and factor loadings, indicating that the specific factor is inseparable from the general factor. In such situations, the offending specific factors should be eliminated and their indicators allowed to load only onto the general factor. Second, like any SEM, bifactor models require sufficiently large sample sizes to provide stable parameter estimates. Sample requirements depend on the degree of communality in the indicators and factor overdetermination (i.e., the degree to which each factor shows strong loadings on multiple indicators; MacCallum, Widaman, Zhang, & Hong, Reference MacCallum, Widaman, Zhang and Hong1999). A concern in bifactor models is that if DGF saturation is large, specific factor loadings will be too weak to provide stable estimates of external relations without very large sample sizes. In such cases, impact of the apparently minor specific factors would not be large enough to justify the costs of trying to measure them reliably.

Comparison of Implications

Results in Table 1 show how choices in data analysis can dominate interpretations of constructs’ relations with external variables in the presence of a DGF. On the basis of correlation analysis, researchers would interpret that both task performance and contextual performance have weak relations with overall satisfaction and satisfaction with work itself, whereas contextual performance is also related to pay and supervisor satisfaction. On the basis of multiple regression, researchers would conclude that both performance variables are associated with overall satisfaction, but task performance is additionally related to satisfaction with the work itself, whereas contextual performance is related to supervisor satisfaction. On the basis of the results of composite correlations and general factor SEM, researchers would conclude that overall satisfaction has little effect on performance. Finally, only the bifactor model was able to separate individuals’ overall evaluations of their job from their beliefs about specific job components, revealing that only the DGF demonstrates a positive relation with performance. All of these are different results that lead to different implications for both theory and applied action.

Presented with the multiple regression results, a theorist might conclude that helping behavior (i.e., contextual performance) stems from good interpersonal relationships at work, whereas task performance is a result of work tasks being intrinsically motivating. Presented with a composite correlation or general factor SEM, a practitioner might decide that improving employee satisfaction would have little impact on tangible organizational outcomes. Critically, by failing to properly account for the DGF using their chosen data analytic method, both conclusions would be based on misleading results and would be wrong. In truth, according to results of the bifactor analysis, task and contextual performance exhibit very similar patterns of relations with the predictors. Most important, for the domain of job satisfaction, only the DGF has a stable positive relation with performance, which is moderate-to-large in magnitude.

Conclusion

DGFs are present in measures of nearly all multidimensional constructs in psychological and organizational research (Ree et al., Reference Ree, Carretta and Teachout2015). Interpretations of general factors vary widely. Some DGFs are artifactual (Conway & Lance, Reference Conway and Lance2010), others are formative composites without independent psychological meaning (e.g., overall job performance; Campbell & Wiernik, Reference Campbell and Wiernik2015), and still others have substantive meaning (Chen et al., Reference Chen, Hayes, Carver, Laurenceau and Zhang2012). However, in many cases, data analytic choices will strongly influence interpretations of DGFs. Determining which interpretation is appropriate for a particular DGF requires not only partitioning the variance within a psychological measure, but also establishing the unique nomological networks of the general and specific factors. Methods that inappropriately conflate DGF and specific factor variance will distort substantive results, causing harm to both theory and practice. In the presence of a DGF, it is essential that researchers use appropriate analytic methods to ensure a valid interpretation of findings.

References

Boker, S. M., Neale, M. C., Maes, H. H., Wilde, M. J., Spiegel, M., Brick, T. R., . . . Team OpenMx. (2015). OpenMx 2.0 user guide (Release No. 2.0.14157). Charlottesville, VA: University of Virginia. Retrieved from http://openmx.psyc.virginia.edu/documentationGoogle Scholar
Campbell, J. P., & Wiernik, B. M. (2015). The modeling and assessment of work performance. Annual Review of Organizational Psychology and Organizational Behavior, 2, 4774. http://doi.org/10.1146/annurev-orgpsych-032414-111427CrossRefGoogle Scholar
Chen, F. F., Hayes, A., Carver, C. S., Laurenceau, J.-P., & Zhang, Z. (2012). Modeling general and specific variance in multifaceted constructs: A comparison of the bifactor model to other approaches. Journal of Personality, 80, 219251. http://doi.org/10/d6ht4bGoogle Scholar
Conway, J. M., & Lance, C. E. (2010). What reviewers should expect from authors regarding common method bias in organizational research. Journal of Business and Psychology, 25, 325334. http://doi.org/10.1007/s10869-010-9181-6Google Scholar
Edwards, B. D., Bell, S. T., Arthur, W. Jr., & Decuir, A. D. (2008). Relationships between facets of job satisfaction and task and contextual performance. Applied Psychology, 57, 441465. http://doi.org/10.1111/j.1464-0597.2008.00328.xGoogle Scholar
Kinicki, A. J., McKee-Ryan, F. M., Schriesheim, C. A., & Carson, K. P. (2002). Assessing the construct validity of the Job Descriptive Index: A review and meta-analysis. Journal of Applied Psychology, 87, 1432. http://doi.org/10.1037/0021-9010.87.1.14Google Scholar
MacCallum, R. C., Widaman, K. F., Zhang, S., & Hong, S. (1999). Sample size in factor analysis. Psychological Methods, 4, 8499. http://doi.org/10.1037/1082-989X.4.1.84Google Scholar
McAbee, S. T., Oswald, F. L., & Connelly, B. S. (2014). Bifactor models of personality and college student performance: A broad versus narrow view. European Journal of Personality, 28, 604619. http://doi.org/10.1002/per.1975Google Scholar
Ree, M. J., Carretta, T. R., & Teachout, M. S. (2015). Pervasiveness of dominant general factors in organizational measurement. Industrial and Organizational Psychology: Perspectives on Science and Practice, 8 (3), 409427.CrossRefGoogle Scholar
Reise, S. P. (2012). The rediscovery of bifactor measurement models. Multivariate Behavioral Research, 47, 667696. http://doi.org/10.1080/00273171.2012.715555CrossRefGoogle ScholarPubMed
Salgado, J. F., Moscoso, S., & Berges, A. (2013). Conscientiousness, its facets, and the prediction of job performance ratings: Evidence against the narrow measures. International Journal of Selection and Assessment, 21, 7484. http://doi.org/10/4qwGoogle Scholar
Wanous, J. P., Reichers, A. E., & Hudy, M. J. (1997). Overall job satisfaction: How good are single-item measures? Journal of Applied Psychology, 82, 247252. http://doi.org/10.1037/0021-9010.82.2.247Google Scholar
Figure 0

Table 1. Results From Four Common Data Analytic Methods for Job Satisfaction

Figure 1

Figure 1. Standardized loadings for a bifactor model of the Edwards et al. (2008) job satisfaction measures with 95% bootstrapped confidence intervals.