Crossref Citations
This article has been cited by the following publications. This list is generated based on data provided by
Crossref.
2017.
Open Peer Commentary and Author'S Response.
European Journal of Personality,
Vol. 31,
Issue. 5,
p.
441.
Jackson, Duncan J. R.
LoPilato, Alexander C.
Hughes, Dan
Guenole, Nigel
and
Shalfrooshan, Ali
2017.
The internal structure of situational judgement tests reflects candidate main effects: Not dimensions or situations.
Journal of Occupational and Organizational Psychology,
Vol. 90,
Issue. 1,
p.
1.
Lievens, Filip
2017.
Construct-Driven SJTs: Toward an Agenda for Future Research.
International Journal of Testing,
Vol. 17,
Issue. 3,
p.
269.
Motowidlo, Stephan J.
Lievens, Filip
and
Ghosh, Kamalika
2018.
Prosocial implicit trait policies underlie performance on different situational judgment tests with interpersonal content.
Human Performance,
Vol. 31,
Issue. 4,
p.
238.
Kaminski, Katarina
Felfe, Jörg
Schäpers, Philipp
and
Krumm, Stefan
2019.
A closer look at response options: Is judgment in situational judgment tests a function of the desirability of response options?.
International Journal of Selection and Assessment,
Vol. 27,
Issue. 1,
p.
72.
Oostrom, Janneke K.
de Vries, Reinout E.
and
de Wit, Mariska
2019.
Development and validation of a HEXACO situational judgment test.
Human Performance,
Vol. 32,
Issue. 1,
p.
1.
Schäpers, Philipp
Freudenstein, Jan-Philipp
Mussel, Patrick
Lievens, Filip
and
Krumm, Stefan
2020.
Effects of situation descriptions on the construct-related validity of construct-driven situational judgment tests.
Journal of Research in Personality,
Vol. 87,
Issue. ,
p.
103963.
Freudenstein, Jan‐Philipp
Schäpers, Philipp
Roemer, Lena
Mussel, Patrick
and
Krumm, Stefan
2020.
Is it all in the eye of the beholder? The importance of situation construal for situational judgment test performance.
Personnel Psychology,
Vol. 73,
Issue. 4,
p.
669.
Wolcott, Michael D.
Lobczowski, Nikki G.
Zeeman, Jacqueline M.
and
McLaughlin, Jacqueline E.
2020.
Situational judgment test validity: an exploratory model of the participant response process using cognitive and think-aloud interviews.
BMC Medical Education,
Vol. 20,
Issue. 1,
Klassen, Robert M.
Kim, Lisa E.
Rushby, Jade V.
and
Bardach, Lisa
2020.
Can we improve how we screen applicants for initial teacher education?.
Teaching and Teacher Education,
Vol. 87,
Issue. ,
p.
102949.
Tiffin, Paul A.
Paton, Lewis W.
O'Mara, Deborah
MacCann, Carolyn
Lang, Jonas W. B.
and
Lievens, Filip
2020.
Situational judgement tests for selection: Traditional vs construct‐driven approaches.
Medical Education,
Vol. 54,
Issue. 2,
p.
105.
Schäpers, Philipp
Lievens, Filip
Freudenstein, Jan‐Philipp
Hüffmeier, Joachim
König, Cornelius J.
and
Krumm, Stefan
2020.
Removing situation descriptions from situational judgment test items: Does the impact differ for video‐based versus text‐based formats?.
Journal of Occupational and Organizational Psychology,
Vol. 93,
Issue. 2,
p.
472.
Cullen, Michael J.
Zhang, Charlene
Marcus-Blank, Brittany
Braman, Jonathan P.
Tiryaki, Ezgi
Konia, Mojca
Hunt, Matthew A.
Lee, Michael S.
Van Heest, Ann
Englander, Robert
Sackett, Paul R.
and
Andrews, John S.
2020.
Improving Our Ability to Predict Resident Applicant Performance: Validity Evidence for a Situational Judgment Test.
Teaching and Learning in Medicine,
Vol. 32,
Issue. 5,
p.
508.
Shiner, Rebecca L.
Soto, Christopher J.
and
De Fruyt, Filip
2021.
Personality Assessment of Children and Adolescents.
Annual Review of Developmental Psychology,
Vol. 3,
Issue. 1,
p.
113.
Martin-Raugh, Michelle P.
and
Kell, Harrison J.
2021.
A process model of situational judgment test responding.
Human Resource Management Review,
Vol. 31,
Issue. 2,
p.
100731.
Pavlidou, Kyriaki
and
Alevriadou, Anastasia
2022.
An Assessment of General and Special Education Teachers’ and Students’ Interpersonal Competences and Its Relationship to Burnout.
International Journal of Disability, Development and Education,
Vol. 69,
Issue. 3,
p.
1080.
Olinover, Mili
Gidron, Maor
Yarmolovsky, Jessica
and
Geva, Ronny
2022.
Strategies for improving decision making of leaders with ADHD and without ADHD in combat military context.
The Leadership Quarterly,
Vol. 33,
Issue. 6,
p.
101575.
Golubovich, Juliya
Lake, Christopher J.
Seybert, Jacob
and
Anguiano‐Carrasco, Cristina
2022.
Do situational characteristics affect the validity of personality situational judgment items?.
International Journal of Selection and Assessment,
Vol. 30,
Issue. 3,
p.
427.
Wang, Dan
Oostrom, Janneke K.
and
Schollaert, Eveline
2023.
The importance of situation evaluation and the ability to identify criteria in a construct-driven situational judgment test.
Personality and Individual Differences,
Vol. 208,
Issue. ,
p.
112182.
Freudenstein, Jan-Philipp
Mussel, Patrick
and
Krumm, Stefan
2023.
On the Construct-Related Validity of Implicit Trait Policies.
European Journal of Personality,
Vol. 37,
Issue. 1,
p.
113.
The construct validity of situational judgment tests (SJTs) is a “hot mess.” The suggestions of Lievens and Motowidlo (Reference Lievens and Motowidlo2016) concerning a strategy to make the constructs assessed by an SJT more “clear and explicit” (p. 5) are worthy of serious consideration. In this commentary, we highlight two challenges that will likely need to be addressed before one can develop SJTs with clear and explicit constructs. We also offer critiques of four positions presented by Lievens and Motowidlo that are not well supported by evidence.
Challenges to Establishing SJT Construct Clarity
The two main challenges likely to complicate the effort of developing SJTs with clearly defined constructs are that (a) SJT items are typically heterogeneous at the item level and (b) SJT scales will typically not show discriminant validity.
SJT Items Are Typically Heterogeneous at the Item Level
As illustrated by McDaniel and Whetzel (Reference McDaniel and Whetzel2005), SJT items are heterogeneous at the item level in that they have correlations with constructs that are not related to each other. For example, an item may have meaningfully large correlations with both general cognitive ability and the personality trait of agreeableness. This makes it difficult to obtain an interpretable factor structure that could be used to determine the constructs measured. Indeed, very few interpretable factor analyses have been reported (for an exception, see Legree, Heffner, Psotka, Martin, & Medsker, Reference Legree, Heffner, Psotka, Martin and Medsker2003). In brief, evidence supporting the construct validity of SJTs is unlikely to come through exploratory or confirmatory factor analyses. Instead, alternative strategies may need to be used to establish the construct validity of an SJT.
SJT Scales Will Typically Not Show Discriminant Validity
Consistent with the typical finding of uninterpretable factors, it is improbable that SJT scales will be able to show discriminant validity. As an example, if one designs SJT scales to measure the Big Five personality traits, the resulting five SJT scales will tend to have much larger magnitude correlations with each other than would be desirable. Furthermore, these correlations will tend to be larger in magnitude than are those found in scales using personality items.
Position Critiques
In addition to the two challenges noted above, we also offer four critiques of Lievens and Motowidlo's positions. We suggest that (a) the inclusion of situational scenarios may help to reduce ambiguity in response options, (b) an alternate view of knowledge acquisition could account for overlap between job-related knowledge and the general knowledge domain, (c) compound traits may lead to unnecessary construct proliferation, and (d) there is no documented empirical evidence that single-item SJTs are more time and cost effective to develop.
Situational Scenarios May Help To Reduce Ambiguity in Response Options
In regard to the authors’ assertion that recent evidence from Krumm et al. (Reference Krumm, Lievens, Hüffmeier, Lipnevich, Bendels and Hertel2015) showed that situational scenarios are not necessary for high performance on SJTs, we argue that situational scenarios can help reduce ambiguities in item responses. McDaniel, Psotka, Legree, Yost, and Weekley (Reference McDaniel, Psotka, Legree, Yost and Weekley2011) have argued that SJT items vary in ambiguity such that the respondent may need to make specific assumptions in order to respond. Such items are associated with near zero validity. Consider this single-response SJT item: “You complete the work assigned to you by two different supervisors, both of whom consider their work to have priority, in the order it was assigned.” Depending on unstated aspects of the situation, doing work in the order in which it was assigned could be either an effective or an ineffective behavior. An SJT that incorporates a scenario can provide needed context to reduce the ambiguity of the response and improve the validity of the item. Thus, Lievens and Motowidlo's suggestion that situational scenarios are not necessary is unlikely to generalize across many SJTs and situations. More research is needed to determine the conditions under which situational scenarios are not required or necessary.
An Alternative View of Knowledge Acquisition
The authors of the focal article also claim that the findings from Krumm et al. (Reference Krumm, Lievens, Hüffmeier, Lipnevich, Bendels and Hertel2015) support the reconceptualization of SJTs as measuring general domain knowledge. However, there can be considerable overlap between job specific knowledge and general domain knowledge. For instance, the claim that “job-specific knowledge can be learned only through exposure to that job or jobs like it” (Lievens & Motowidlo, p. 8) is not necessarily correct. Job specific knowledge can be obtained through formal education and training. For example, a doctoral student can learn about item response theory (a job specific knowledge for a test developer) in graduate school, retain this knowledge, and apply it when employed as test developer. Furthermore, the authors’ assertion that “general domain knowledge is . . . not acquired from specific job experiences” (Lievens & Motowidlo, p. 8) is also likely to be inaccurate. Consider an adolescent, employed in a fast food restaurant, who arrives to work late. Upon arrival, the supervisor counsels the adolescent regarding the inappropriateness of being late. Due to this situation, the adolescent has obtained general knowledge about the value of being on time for work. Therefore, the authors’ claim may not be correct. Any (re-) conceptualization of SJTs should account for the strong likelihood of considerable overlap in the acquisition of job specific and general domain knowledge.
Compound Traits Such as Prosocial Action May Contribute to Construct Proliferation
The authors suggest that their approach can be used to measure compound traits. Although SJTs can be developed to measure compound traits, we caution against this. Compound traits likely contribute to the construct proliferation that plagues the industrial–organizational psychology and management literature like an ever-expanding clump of fungus devouring our discipline (Le, Schmidt, Harter, & Lauver, Reference Le, Schmidt, Harter and Lauver2010; Schwab, Reference Schwab, Staw and Cummings1980). In addition, designing an SJT to measure a compound trait ultimately runs contrary to the authors’ stated goal, namely, to develop SJTs with “clear and explicit constructs” (Lievens & Motowidlo, p. 5), because compound traits are inherently multidimensional. Therefore, we encourage the reevaluation of the suggestion that SJTs should be developed to assess compound traits.
Undocumented Claims Concerning Time and Cost Efficiency of Single-Response SJT Items
Although we have nothing against single-response SJTs, the authors of the focal article assert, without support, that single-response “item development . . . is further simplified and made more efficient” (Lievens & Motowidlo, p. 17). We disagree. For example, with a Likert rating format (e.g., “rate each of the responses using the 1–7 scale of effectiveness”), each response associated with a scenario is a scorable item. Thus, for a scenario with five response options, one can obtain five scorable items with one scenario. With a single-response item, one needs one scenario for each scorable item. We suggest that our difference of opinion with the focal article authors would best be addressed empirically.
Conclusion
Taken together, we concur with Lievens and Motowidlo that the construct validity of SJTs could benefit greatly from additional research attention. In fact, we believe that it is currently a “hot mess” without much theoretical or empirical guidance. We also agree with several of the authors’ suggestions. However, we see some challenges in the strategies offered by Lievens and Motowidlo and disagree with several of their assertions.