Crossref Citations
This article has been cited by the following publications. This list is generated based on data provided by
Crossref.
Dilchert, Stephan
2017.
Future of research published in the International Journal of Selection and Assessment: Incoming editor's perspective.
International Journal of Selection and Assessment,
Vol. 25,
Issue. 4,
p.
416.
Dahlke, Jeffrey A.
and
Wiernik, Brenton M.
2017.
Nei, Kimberly S.
Foster, Jeff L.
Ness, Alisha M.
and
Nei, Darin S.
2018.
Rule breakers and attention seekers: Personality predictors of integrity and accountability in leaders.
International Journal of Selection and Assessment,
Vol. 26,
Issue. 1,
p.
17.
Wille, Bart
Wiernik, Brenton M.
Vergauwe, Jasmine
Vrijdags, Amelie
and
Trbovic, Nikola
2018.
Personality characteristics of male and female executives: Distinct pathways to success?.
Journal of Vocational Behavior,
Vol. 106,
Issue. ,
p.
220.
Elson, Malte
2019.
Examining Psychological Science Through Systematic Meta-Method Analysis: A Call for Research.
Advances in Methods and Practices in Psychological Science,
Vol. 2,
Issue. 4,
p.
350.
Harari, Michael B.
Parola, Heather R.
Hartwell, Christopher J.
and
Riegelman, Amy
2020.
Literature searches in systematic reviews and meta-analyses: A review, evaluation, and recommendations.
Journal of Vocational Behavior,
Vol. 118,
Issue. ,
p.
103377.
Rudolph, Cort W.
Chang, Caelen K.
Rauvola, Rachel S.
and
Zacher, Hannes
2020.
Meta-analysis in vocational behavior: A systematic review and recommendations for best practices.
Journal of Vocational Behavior,
Vol. 118,
Issue. ,
p.
103397.
Stojanov, Ana
and
Halberstadt, Jamin
2020.
Does lack of control lead to conspiracy beliefs? A meta‐analysis.
European Journal of Social Psychology,
Vol. 50,
Issue. 5,
p.
955.
Wollast, Robin
Riemer, Abigail R.
Sarda, Elisa
Wiernik, Brenton M.
and
Klein, Olivier
2020.
How Self-Compassion Moderates the Relation Between Body Surveillance and Body Shame Among Men and Women.
Mindfulness,
Vol. 11,
Issue. 10,
p.
2298.
Dahlke, Jeffrey A.
and
Wiernik, Brenton M.
2020.
Not Restricted to Selection Research: Accounting for Indirect Range Restriction in Organizational Research.
Organizational Research Methods,
Vol. 23,
Issue. 4,
p.
717.
Zhang, Charlene
and
Kuncel, Nathan R.
2020.
Moving Beyond the Brag Sheet: A Meta‐Analysis of Biodata Measures Predicting Student Outcomes.
Educational Measurement: Issues and Practice,
Vol. 39,
Issue. 3,
p.
106.
Okan, Mehmet
Elmadag, Ayse Banu
and
İdemen, Elif
2021.
Frontline employee age and customer mistreatment: a meta-analysis.
Journal of Services Marketing,
Vol. 35,
Issue. 1,
p.
98.
Harari, Michael B.
McCombs, Kate
and
Wiernik, Brenton M.
2021.
Movement Capital, RAW model, or circumstances? A meta-analysis of perceived employability predictors.
Journal of Vocational Behavior,
Vol. 131,
Issue. ,
p.
103657.
Kleine, Anne-Kathrin
Schmitt, Antje
and
Wisse, Barbara
2021.
Students' career exploration: A meta-analysis.
Journal of Vocational Behavior,
Vol. 131,
Issue. ,
p.
103645.
Tanim, Shakhawat H.
Wiernik, Brenton M.
Reader, Steven
and
Hu, Yujie
2022.
Predictors of hurricane evacuation decisions: A meta-analysis.
Journal of Environmental Psychology,
Vol. 79,
Issue. ,
p.
101742.
La Rue, Crystal J.
Haslam, Catherine
and
Steffens, Niklas K.
2022.
A meta-analysis of retirement adjustment predictors.
Journal of Vocational Behavior,
Vol. 136,
Issue. ,
p.
103723.
Kubicek, Bettina
Uhlig, Lars
Hülsheger, Ute R.
Korunka, Christian
and
Prem, Roman
2023.
Are all challenge stressors beneficial for learning? A meta-analytical assessment of differential effects of workload and cognitive demands.
Work & Stress,
Vol. 37,
Issue. 3,
p.
269.
Morris, Scott B.
2023.
Meta-Analysis in Organizational Research: A Guide to Methodological Options.
Annual Review of Organizational Psychology and Organizational Behavior,
Vol. 10,
Issue. 1,
p.
225.
Fillon, Adrien A.
Girandola, Fabien
Bonnardel, Nathalie
Kenworthy, Jared B.
and
Souchet, Lionel
2024.
Stage 2 Registered Report Personal factors and group creative outcomes: A correlational meta-analysis.
F1000Research,
Vol. 13,
Issue. ,
p.
904.
Traylor, Allison M.
Dinh, Julie V.
Ng, Linnea C.
Reyes, Denise L.
Cheng, Shannon K.
Croitoru, Natalie C.
and
Salas, Eduardo
2024.
It's about the process, not the product: A meta-analytic investigation of team demographic diversity and processes.
Organizational Psychology Review,
Vol. 14,
Issue. 3,
p.
478.
Generalization in meta-analyses is not a dichotomous decision (typically encountered in papers using the Q test for homogeneity, the 75% rule, or null hypothesis tests). Inattention to effect size variability in meta-analyses may stem from a lack of guidelines for interpreting credibility intervals. In this commentary, we describe two methods for making practical interpretations and determining whether a particular SDρ represents a meaningful level of variability.
Normative Interpretation
Typically, meta-analysts evaluate the magnitude of heterogeneity in effect sizes by focusing on the face value of the true variability of the effect size distribution (SDρ) or the width of its associated credibility interval (cf. “the .30-unit width of the interval”; Tett, Hundley, & Christiansen, Reference Tett, Hundley and Christiansen2017, p. 8). This approach needs refinement, however, as correct interpretation of a credibility interval depends not only on its width but also on the range of correlations that it spans. Correlations in applied psychology are not uniformly distributed, and a number of reviews have summarized distributions of effect sizes (e.g., Bosco, Aguinis, Singh, Field, & Pierce, Reference Bosco, Aguinis, Singh, Field and Pierce2015; Gignac & Szodorai, Reference Gignac and Szodorai2016; Hemphill, Reference Hemphill2003). Among them, the most comprehensive review is that of Paterson, Harms, Steel, and Credé (Reference Paterson, Harms, Steel and Credé2016), which integrates 30 years of meta-analyses of micro-level variables in management to develop empirical benchmarks for observed and corrected correlations. An advantage of this approach is that, by constructing distributions from meta-analyses, upward biasing effects of sampling error on effect size variability are controlled.
Paterson et al. (Reference Paterson, Harms, Steel and Credé2016, p. 76) report a general distribution of corrected correlations with quartiles of .15, .25, and .39. According to these benchmarks, the distribution of correlations in applied psychology is positively skewed, with most correlations falling between .15 and .40. A consequence of this nonuniform distribution is that true variability is more substantively meaningful for some correlation ranges than for others. For example, a .20-unit wide interval from .08 to .28 reflects a range from comparatively negligible (<25th percentile of the empirical distribution) to moderate (between 50th and 75th percentiles) correlation values, whereas, the endpoints of a .20-unit wide interval from .55 to .75 would both be considered “very large” (>90th percentile). Variability is more meaningful, both for theoretical interpretations and practical applications (e.g., incremental validity), when it occurs near the center of the empirical distribution of effects.
An alternative to face-value interpretations of SDρ is comparing ρ and credibility interval endpoints to a relevant empirical distribution of effect sizes. Appropriate questions from this perspective include the following: What percentiles of the empirical distribution do the endpoints of the interval correspond to? Do endpoints reflect substantively different parts of the distribution, relative to the research question? What is the percentile distance between top and bottom bounds of the interval? If the credibility interval spans a wide range of the empirical distribution, then there may be value in searching for moderators. By comparison, when an interval covers a narrow slice of the empirical distribution, any potential moderators are unlikely to have a strong impact on theoretical interpretations or practical decisions.
Table 1 illustrates hypothetical combinations of ρ and SDρ values, along with corresponding credibility intervals, percentile values for Paterson et al.’s (Reference Paterson, Harms, Steel and Credé2016) overall empirical distribution for ρ, and the percentile distance between credibility interval endpoints. Several patterns are clear from these values. First, for ρ < .40, SDρ values between .07 and .11 translate to credibility intervals that cover major portions of the distribution of correlations found in industrial and organizational (I-O) psychology research. For larger correlations, however, even sizable SDρ values represent increasingly small slices of the distribution (e.g., endpoints of a credibility interval for ρ = .55, SDρ = .09 are only 21 percentile points apart). For ρ ≥ .60, both bounds of all credibility intervals examined fall in the upper 25% of applied effects. For effects this large, even ostensibly large variability is unlikely to alter substantive conclusions about the strength of relations between constructs.
Table 1. I-O Research Correlation Percentiles for Credibility Intervals
Note. Percentiles based on Paterson et al.’s (Reference Paterson, Harms, Steel and Credé2016) overall distribution of corrected correlations; bars in percentile columns reflect in which quarter of the correlation distribution values fall (white bars with border indicate that one bound of the interval crosses zero); percentile (%ile) distance range plots scaled to minimum of 0 and maximu m of 100; Cred. Int. = credibility interval; CV = credibility value.
Conversely, comparatively small variability can be substantively meaningful for some values of ρ. An SDρ = .03 is often regarded as negligible, and, indeed, for ρ ≤ .15 or ρ ≥ .35, this variability reflects a minor portion of the empirical distribution. However, for correlations between these values, this small SDρ corresponds to credibility intervals that span a wide section of the effect size distribution. Indeed, for a median correlation (ρ = .25), the credibility interval for SDρ = .03 spans 30% of empirical effects.
In sum, there is more substantive value for explaining variability and identifying potential moderators if the credibility interval covers a wide range of the empirical effect sizes. For SDρ values examined here, the largest meaningful variability occurs for |ρ| = .10 to .45. By contrast, variability for larger correlations will not affect central conclusions.
Table 2 presents the percentile ranges of credibility intervals for a sample of published meta-analyses reporting relations to Conscientiousness (based on Wilmot, Reference Wilmot2017). Consistent with the hypothetical credibility intervals in Table 1, the empirical credibility intervals cover the widest portion of the empirical effect size distribution (i.e., have the largest percentile difference) for correlations ranging from .10 to .30. Credibility intervals for larger correlations tend to cover less of the empirical distribution, even for relatively large SDρ values. For larger correlations, there is less value in exploring for moderators, as theoretical conclusions about the impact of Conscientiousness are unlikely to change for different values (e.g., whether Conscientiousness correlates ρ = .45 or .60 with Quality of Life, researchers would still conclude this relationship is very large). An extreme example is the relation of Conscientiousness to Grit. The correlation has a relatively large SDρ (.07), but the mean effect (ρ = .84) is so large that the entire credibility interval is above the 99th percentile of applied effects.Footnote 1 Thus, moderator searches will be more meaningful for criteria with ρ located in the center of the distribution.
Table 2. Percentile Interpretations of Selected Published Credibility Intervals for Conscientiousness
Note. Table columns and color-coding correspond to those in Table 1; †reverse-coded; OCB = organizational citizenship behavior; CWB = counterproductive work behavior; all correlations corrected for unreliability in both variables; tabulated data based on Wilmot (Reference Wilmot2017).
Objective Interpretation
When dependent variables are expressed in units with meaningful, objective outcomes for individuals or organizations (e.g., salary, turnover rates, production costs), credibility intervals can also be interpreted in terms of their practical implications. For these criteria, variability can be translated into real-world consequences of correlations at each end of the interval. From this perspective, important questions include the following: Would applied decisions (e.g., to use a test in employee selection) change for the highest versus lowest interval values? Would conclusions about the importance of a variable for a real-world outcome change across the range of these correlations?
As an example, using the correlation with salary in Table 2 and the U. S. Census Bureau–reported SD of annual income for adults working ≥ 40 hours per week of $33,647, the impact of a 1 SD increase in Conscientiousness on individual salary varies from $7,066 for the upper bound value to negative $2,692 for the lower bound. This nearly $10,000 swing (and reversal of the effect's direction) clearly has practical implications for individuals and warrants investigation of moderators. Conversely, for turnover, assuming a base rate of 50% turnover and a selection ratio of .30, the lower and upper credibility values reflect a difference in reduced turnover rates of only 3%, which is less likely to substantively alter organizational decision making.
Conclusions and Caveats
Dichotomous conclusions about homogeneity or generalizability are not useful or appropriate, just as dichotomous conclusions about effects based on statistical significance tests lead to erroneous conclusions (see Schmidt, Reference Schmidt2008). Normative and objective interpretations of credibility intervals can help producers and consumers of meta-analyses to grow beyond reductionist descriptions of variability and simplistic interpretations of SDρ. Accounting for variability matters more when the credibility interval spans a sufficiently wide portion of the empirical distribution of effects. Moreover, values of SDρ that translate into meaningful variability depend on the position of ρ in the empirical distribution. As a result, proper interpretations rely on accurate estimates of both ρ and SDρ. For the latter, precise estimates of SDρ require larger k values than are typical in I-O psychology meta-analyses (Schmidt, Reference Schmidt2008; Steel, Kammeyer-Mueller, & Paterson, Reference Steel, Kammeyer-Mueller and Paterson2015), as well as the ability to fully correct for statistical artefacts using robust artefact estimates.Footnote 2 To increase precision, future meta-analyses may consider placing confidence intervals around SDρ and/or using a Bayesian estimator with priors based on relevant variables. As meta-analytic research progresses and informs practice, accurately estimating true variability and properly characterizing its implications, is paramount.