Crossref Citations
This article has been cited by the following publications. This list is generated based on data provided by
Crossref.
Clark, Benjamin Y.
and
Brudney, Jeffrey L.
2014.
Representation and Beyond: Coproduction and the New Information Technology.
SSRN Electronic Journal,
Jia, Ronnie
Reich, Blaize Horner
and
Jia, Heather H.
2016.
A commentary on: “Creating agile organizations through IT: The influence of IT service climate on IT service quality and IT agility”.
The Journal of Strategic Information Systems,
Vol. 25,
Issue. 3,
p.
227.
Keith, Melissa G.
Tay, Louis
and
Harms, Peter D.
2017.
Systems Perspective of Amazon Mechanical Turk for Organizational Research: Review and Recommendations.
Frontiers in Psychology,
Vol. 8,
Issue. ,
Lemoine, J.
Kmiec, R.
and
Roland-Lévy, C.
2017.
Characterization of controlled gamblers and pathological gamblers using the social representation theory.
European Review of Applied Psychology,
Vol. 67,
Issue. 1,
p.
13.
Kees, Jeremy
Berry, Christopher
Burton, Scot
and
Sheehan, Kim
2017.
An Analysis of Data Quality: Professional Panels, Student Subject Pools, and Amazon's Mechanical Turk.
Journal of Advertising,
Vol. 46,
Issue. 1,
p.
141.
Cheung, Janelle H.
Burns, Deanna K.
Sinclair, Robert R.
and
Sliter, Michael
2017.
Amazon Mechanical Turk in Organizational Psychology: An Evaluation and Practical Recommendations.
Journal of Business and Psychology,
Vol. 32,
Issue. 4,
p.
347.
Martin-Lacroux, Christelle
and
Lacroux, Alain
2017.
Do Employers Forgive Applicants’ Bad Spelling in Résumés?.
Business and Professional Communication Quarterly,
Vol. 80,
Issue. 3,
p.
321.
Lacroux, Alain
and
Martin-Lacroux, Christelle
2017.
Quelle efficacité pour le CV anonyme ? Les leçons d’une étude expérimentale.
Revue de gestion des ressources humaines,
Vol. N° 104,
Issue. 2,
p.
61.
Gleibs, Ilka H.
2017.
Are all “research fields” equal? Rethinking practice for the use of data from crowdsourcing market places.
Behavior Research Methods,
Vol. 49,
Issue. 4,
p.
1333.
O'Hare, Sheila
and
Erdelez, Sanda
2017.
Legal information acquisition by the public: The role of personal and environmental factors.
Proceedings of the Association for Information Science and Technology,
Vol. 54,
Issue. 1,
p.
298.
Clark, Benjamin Y.
and
Brudney, Jeffrey L.
2018.
Citizen Representation in City Government-Driven Crowdsourcing.
Computer Supported Cooperative Work (CSCW),
Vol. 27,
Issue. 3-6,
p.
1153.
Mueller, J. Tom
Mowen, Andrew J.
and
Graefe, Alan R.
2018.
We Aren’t So Different After All: Differences and Similarities Between Political Affiliation and Issues of Park Use, Management, and Privatization.
Leisure Sciences,
Vol. 40,
Issue. 7,
p.
735.
Nguyen, Diep T. N.
Teo, Stephen T. T.
Pick, David
and
Jemai, Mohamed
2018.
Cynicism about Change, Work Engagement, and Job Satisfaction of Public Sector Nurses.
Australian Journal of Public Administration,
Vol. 77,
Issue. 2,
p.
172.
Keith, Melissa G.
Harms, Peter
and
Tay, Louis
2019.
Mechanical Turk and the gig economy: exploring differences between gig workers.
Journal of Managerial Psychology,
Vol. 34,
Issue. 4,
p.
286.
Cerri, Jacopo
Testa, Francesco
Rizzi, Francesco
and
Frey, Marco
2019.
Factorial surveys reveal social desirability bias over self-reported organic fruit consumption.
British Food Journal,
Vol. 121,
Issue. 4,
p.
897.
Morean, Meghan E.
and
Lederman, Izzy R.
2019.
Prevalence and correlates of medical cannabis patients' use of cannabis for recreational purposes.
Addictive Behaviors,
Vol. 93,
Issue. ,
p.
233.
Neville, Lukas
and
Fisk, Glenda M.
2019.
Getting to Excess: Psychological Entitlement and Negotiation Attitudes.
Journal of Business and Psychology,
Vol. 34,
Issue. 4,
p.
555.
Behrend, Tara S.
and
Ravid, Daniel M.
2019.
The Cambridge Handbook of Technology and Employee Behavior.
p.
790.
Lacroux, Alain
and
Matin-Lacroux, Christelle
2019.
Spelling Error Perception During Personnel Selection: A Mixed Methods Study of Recruiters’ Judgments Using Lexicometric Analysis.
International Journal of Multiple Research Approaches,
Vol. 11,
Issue. 2,
p.
183.
Stedman, Richard C.
Connelly, Nancy A.
Heberlein, Thomas A.
Decker, Daniel J.
and
Allred, Shorna B.
2019.
The End of the (Research) World As We Know It? Understanding and Coping With Declining Response Rates to Mail Surveys.
Society & Natural Resources,
Vol. 32,
Issue. 10,
p.
1139.
In their focal article, Landers and Behrend (Reference Landers and Behrend2015) propose to reevaluate the legitimacy of using the so-called convenience samples (e.g., crowdsourcing, online panels, and student samples) as compared with traditional organizational samples in industrial–organizational (I-O) psychology research. They suggest that such sampling strategies should not be judged as inappropriate per se but that decisions to accept or reject such samples must be empirically or theoretically justified. I concur with Landers and Behrend's call for a more nuanced view on convenience samples. More precisely, I suggest that we should not “throw the baby out with the bathwater” but rather carefully and empirically examine the advantages and risks associated with using each sampling strategy before classifying it as suitable or not.
In this commentary, I examine and compare original data obtained from the three types of convenience samples highlighted in the focal article: (a) crowdsourcing, (b) online panels with participants recruited by commercial panel providers, and (c) business students. I discuss differences with regards to sample composition (i.e., age, gender, ethnicity, education, and employment status). Moreover, using the example of a measure of competitive worldviews (CWs; Duckitt, Wagner, Du Plessis, & Birum, Reference Duckitt, Wagner, Du Plessis and Birum2002), I examine data quality (i.e., mean differences and range restriction, scale reliability, data normality). I compare data from the original convenience samples to published data in the social and personality psychology literatures obtained with two more traditional (but still convenience) sample types: psychology students and survey data from participants recruited from the general adult population.
Convenience Samples Composition
In the focal article, Landers and Behrend (Reference Landers and Behrend2015) argue that online sampling strategies such as crowdsourcing (e.g., Mechanical Turk) may reduce research overreliance on WEIRD (Western, educated, industrialized, rich, and democratic) samples. Such strategies may be considered as a viable alternative to the traditional convenience samples composed of young and inexperienced college students, for instance when organizational samples are difficult to access. Indeed, Mechanical Turk (MTurk) participants have been described as more diverse than traditional college samples and more representative of the general (U.S.) population (Buhrmester, Kwang, & Gosling, Reference Buhrmester, Kwang and Gosling2011; Paolacci, Chandler, & Ipeirotis, Reference Paolacci, Chandler and Ipeirotis2010). However, there is limited literature discussing the composition of other forms of online panels (e.g., provided by commercial participant recruitment services like Qualtrics or SurveyMonkey), probably because their composition depends on the specific needs or requests of researchers and the screening procedure used by the panel providers (Brandon, Long, Loraas, Mueller-Phillips, & Vansant, Reference Brandon, Long, Loraas, Mueller-Phillips and Vansant2013). For instance, researchers can request samples from specific countries or regions or select panel participants based on age, gender, or education level.
In order to examine the composition of convenience samples further, I use original data obtained from nine independent samples.Footnote 1 More precisely, the data come from two MTurk samples with U.S. participants (Ns = 510 and 488), four samples from Qualtrics online panels (N = 105 from Switzerland, N = 104 from Germany, N = 104 from Greece, and N = 102 from Spain), and three samples with business students (N = 105 from Canada, N = 67 from Germany, and N = 57 from Switzerland). Furthermore, I compare this original data with data from 16 samples from published articles using the CWs measure:Footnote 2 11 samples using psychology students and five samples using a general adult population (i.e., approached in public places in large cities and sent a survey questionnaire). The demographic composition of each sample type is presented in Table 1.
Table 1 Demographics and Competitive Worldviews Data Across Samples
Note. All competitive worldviews (CWs) scores for the original data are based on the original 20-item CWs with a 1–5 Likert scale; CWs scores from published data are mostly based on a sample of items from the original scale using 1–7 or −4–4 Likert scales (transformed into 1–5 scores for comparison); Ethnicity statistics are based on U.S. samples only. Range for the original samples was the observed range from the data, whereas range for the published sample was estimated based on M +/−3SD.
Participants in both the MTurk samples and the online panels were older than were those from the original business student samples (Cohen's d = 0.75, 95% CI [0.6; 0.9], and d = 1.25, 95% CI [1.07; 1.42], respectively; CI = confidence interval) and the published psychology student samples (d = 1.75, 95% CI [1.67; 1.83], and d = 2.41, 95% CI [2.29; 2.52]). They were also much more similar to the general population surveys (d = −0.12, 95% CI [−0.2; −0.04], and d = −0.12, 95% CI [−0.23; −0.01]). Gender parity was almost achieved in all three original convenience sample types, which was not the case with the published samples (both students and the general population surveys). About half of the participants possessed university/college-level education both with MTurk (53%) and the online panel participants (47%, although it varied from country to country). Regarding employment status, both MTurk samples and online panel samples included mostly employed (48% and 60%) but also unemployed (31% and 30%) and student (21% and 10%) participants (although it varied from country to country in the online panels). No information about education or employment status was provided for the general population surveys. Finally, the ethnicity of the U.S. samples was examined. The composition of MTurk samples was diverse, albeit slightly overrepresenting Whites and Asians but underrepresenting Blacks and Hispanics as compared with the U.S. workforce (U.S. Bureau of Labor Statistics, 2014). Interestingly, a similar conclusion can be reached for the ethnic composition of U.S. student samples in published research (i.e., overrepresentation of Asians but underrepresentation of Blacks and Hispanics).
Altogether, these results suggest that MTurk and online panels may not perfectly mirror organizational samples, for instance, because they still incorporate unemployed individuals or students. However, they offer access to samples that in some aspects resemble the working population more than student samples (e.g., older and likely more experienced). It is for this reason that I-O psychology journals are less reluctant toward masters of business administration samples than undergraduate student samples (Landers & Behrend, Reference Landers and Behrend2015). Given the results shown in this commentary, they may want to consider MTurk and online panels as viable alternatives, for the same reason. Moreover, online samples may include a large proportion of employed and unemployed job seekers, which may be particularly relevant for specific research areas (e.g., selection and recruitment, vocational and career adjustment). For instance, 43% of our MTurk and 62% of our online panel participants reported that they were actively looking for jobs at the time of our data collections.
Convenience Samples Data Quality
Landers and Behrend's (Reference Landers and Behrend2015) focal article also discussed the issue of data quality when using convenience samples. For instance, they argue that online samples like MTurk may reduce problems with range restriction on the variables of interest, an issue sometimes associated with organizational samples because employees may have been selected based on this variable (e.g., cognitive ability). Earlier examinations of MTurk have highlighted that high-quality data can be obtained even with low compensation (e.g., Behrend, Sharek, Meade, & Wiebe, Reference Behrend, Sharek, Meade and Wiebe2011; Buhrmester et al., Reference Buhrmester, Kwang and Gosling2011; Crump, McDonnell, & Gureckis, Reference Crump, McDonnell and Gureckis2013). However, those studies have focused their effort on reliability alpha, test–retest reliability, or experimental studies but have overlooked range restriction or data normality. Other online panels (e.g., SurveyMonkey or Qualtrics) have also been described as offering researchers access to data of adequate quality (Brandon et al., Reference Brandon, Long, Loraas, Mueller-Phillips and Vansant2013), but existing investigations have focused on completion rates or correct responses to manipulation check questions.
I provide some additional evidence about the quality of data obtained with MTurk, online panels (Qualtrics) and student samples by examining mean differences, potential range restriction, scale reliability, and data normality (see Table 1). In order to do so, all participants in the nine samples completed the same 20-item CWs scale. CWs capture people's tendency to perceive the world as a competitive jungle characterized by a ruthless struggle to obtain scarce resources (Duckitt et al., Reference Duckitt, Wagner, Du Plessis and Birum2002). CWs represent a pertinent measure to assess data quality because the factor used to screen participants in the nine samples (i.e., a recent experience with selection) is theoretically unrelated to CWs, thus offering a “clean” example of the risks versus the benefits associated with convenience samples. Like with sample composition, I also compare the original convenience samples with published studies using data from psychology student samples or general adult population surveys.
CWs scores for MTurk were higher than were psychology students’ (d = 0.16, 95% CI [0.09; 0.23]) but were similar to business students’ (d = −0.11, 95% CI [−0.25; 0.04]) and the general population's scores. (d = −0.11, 95% CI [−0.19; −0.02]). CWs scores for the online panels were higher than were psychology students’ (d = 0.4, 95% CI [0.3; 0.5]) but were similar to business students’ (d = 0.13, 95% CI [−0.03; 0.3]) and the general population's (d = 0.08, 95% CI [−0.04; 0.18]) scores. MTurk CWs scores were lower than were the online panels’ (d = −0.24, 95% CI [−0.35; −0.12]) score. The theoretical range of CWs scores for the general population was 3.56. Among the various convenience samples, MTurk offered the least range restriction (3.10). Psychology students (2.83), business students (2.80), and online panel data (2.40) offered more range restricted CWs data. The reliability coefficients obtained with all three types of convenience samples for the CWs measure were excellent and equivalent (if not superior) to coefficients of published studies using psychology students or survey data from the general population. Finally, data normality was examined with skewness and kurtosis measures. In all three sample types (even with rather large samples), converted z scores were below the traditional significance threshold (i.e., 1.96), suggesting that samples distribution did not deviate from normality.
Altogether, the scores of all three original convenience samples (i.e., MTurk, online panels, and business students) were similar to those of the general adult population, while psychology student samples obtained lower scores. The issue of range restriction was limited for MTurk, but it appeared to be more important for online panels. Moreover, data obtained with all three sample types (including MTurk and online panels) achieved good reliability. Finally, convenience sample data normality was not an issue. Overall, these results offer support to Landers and Behrend's (Reference Landers and Behrend2015) claims and complement earlier evidence for MTurk (Buhrmester et al., Reference Buhrmester, Kwang and Gosling2011) and online panels (Brandon et al., Reference Brandon, Long, Loraas, Mueller-Phillips and Vansant2013) data quality. Yet, they are based on only one measure and a small number of samples and should be extended to other measures and samples.
Conclusion
This commentary aimed at better understanding the potential value and risks associated with convenience samples. Crowdsourcing (e.g., MTurk) represents a fast and inexpensive sampling method, allowing researchers to collect data from quite large samples in only hours for as cheap as a few cents or dollars per participant (Behrend et al., Reference Behrend, Sharek, Meade and Wiebe2011). Moreover, screening questions can also allow researchers to target participants who possess specific characteristics (e.g., a recent experience with selection in the samples presented here). I present here some additional evidence showing that MTurk can offer access to reliable and high-quality data from participants whose demographics resemble the general labor force more than student samples (e.g., older, more diverse). Online panels like Qualtrics represent another alternative, as they offer access to reliable data that also resemble the general labor force. Moreover, although data collection is likely to be more costly with commercial online panels than with MTurk, such sampling strategies may help researchers to access samples from regions that are currently underrepresented in MTurk (e.g., some European countries).
More research is certainly necessary to better understand how and when crowdsourcing and online panels constitute viable alternatives to organizational samples for I-O psychology research. Yet, evidence is (slowly) accumulating to support Landers and Behrend's (Reference Landers and Behrend2015) call for researchers, reviewers, and editors to not automatically condemn such sampling strategies.