Hostname: page-component-586b7cd67f-dlnhk Total loading time: 0 Render date: 2024-11-27T20:30:09.679Z Has data issue: false hasContentIssue false

Pay Rates and Subject Performance in Social Science Experiments Using Crowdsourced Online Samples

Published online by Cambridge University Press:  15 May 2018

David J. Andersen
Affiliation:
Department of Political Science, Iowa State University, 547 Ross Hall, Ames, IA 50011-1204, USA, e-mail: [email protected]
Richard R. Lau
Affiliation:
Department of Political Science, Rutgers University, 89 George Street, New Brunswick, NJ 08901-1411, USA, e-mail: [email protected]
Rights & Permissions [Opens in a new window]

Abstract

Mechanical Turk has become an important source of subjects for social science experiments, providing a low-cost alternative to the convenience of using undergraduates while avoiding the expense of drawing fully representative samples. However, we know little about how the rates we pay to “Turkers” for participating in social science experiments affects their participation. This study examines subject performance using two experiments – a short survey experiment and a longer dynamic process tracing study of political campaigns – that recruited Turkers at different rates of pay. Looking at demographics and using measures of attention, engagement and evaluation of the candidates, we find no effects of pay rates upon subject recruitment or participation. We conclude by discussing implications and ethical standards of pay.

Type
Research Article
Copyright
Copyright © The Experimental Research Section of the American Political Science Association 2018 

“Crowdsourcing” samples have emerged as a fast, easy, and inexpensive source of subjects for experimental research. In particular, Amazon's Mechanical Turk has become a popular source for quickly and cheaply recruiting large numbers of respondents (Berinsky et al., Reference Berinsky, Huber and Lenz2012; Paolacci et al., Reference Paolacci, Chandler and Ipeirotis2010). “Turkers,” as they are known, are a ready alternative to undergraduates or professionally assembled samples, and offer two major benefits: their availability (Hitlin, Reference Hitlin2016; though also see Stewart et al., Reference Stewart, Ungemach, Harris, Bartels, Newll, Paolacci and Chandler2015) and their inexpensive cost, while still providing a diverse pool of subjects (Huff and Tingley, Reference Huff and Tingley2015; Ipeirotis, Reference Ipeirotis2010; Levay et al., Reference Levay, Freese and Druckman2016).

Determining what to pay subjects on Mechanical Turk can be challenging for two reasons that may risk the quality of the sample recruited. First, different pay rates may attract different participants. Turkers selectively choose which available HITs they will accept, making it possible that the selection process may introduce sample biases (Krupnikov and Levine, Reference Krupnikov and Levine2014). Higher pay rates may attract a different type of worker than lower pay rates, either demographically or along some other factor that might influence subject performance. Second, paying too little in compensation may lead to sub-par subject attention, as participants who decide they are not going to be sufficiently compensated alter their performance (Berinsky et al., Reference Berinsky, Margolis and Sances2016).

For simple tasks with “right” or “wrong” results that the Requester can evaluate, there is an easy mechanism for evaluating subject behavior – rewarding accurate behavior through payment and punishing inaccurate behavior by denying payment. The Requester simply checks on the work as it is returned to make sure that the Worker was indeed paying attention and performing adequately. Turkers know this, and behave accordingly.

As Ho, Slivkins, Suri, and Vaughan describe: “even when standard, unconditional payments are used and no explicit acceptance criteria is specified, workers may behave as if the payments are implicitly performance-based since they believe their work may be rejected if its quality is sufficiently low” (Ho et al., Reference Ho, Slivkins, Suri and Vaughan2015). In such scenarios, different pay rates have been demonstrated to motivate workers to do a greater quantity of work, but not at higher quality (Mason and Watts, Reference Mason and Watts2009). Similarly, several studies have shown that, when work is verifiable based upon accuracy or correctness, pay rates can influence worker behavior positively (Finnerty et al., Reference Finnerty, Kucherbaev, Tranquillini and Convertino2013; Horton and Chilton, Reference Horton and Chilton2010; Ho et al., Reference Ho, Slivkins, Suri and Vaughan2015; Ye et al., Reference Ye, You and Robert2017).

Social scientists should take pause at this, because all of these studies are conditional upon the ability to review subject performance using objective criteria. For example, determining if a subject correctly ordered images, or successfully identified words among a jumble of letters is relatively easy (Mason and Watts, Reference Mason and Watts2009). However, subject performance in social scientific studies tends to lack a strong evaluation component. That is, subjects are asked to behave “normally” and react to the information and stimuli they are provided as they would in the real-world, but without the ability of the experimenter to verify that they are indeed doing so. Behaving “normally” does not clearly indicate a “right” or “wrong” set of behaviors that can be observed. It is exceedingly difficult to determine if a subject is paying attention to an online study (Berinsky et al., Reference Berinsky, Huber and Lenz2012; Berinsky et al., Reference Berinsky, Margolis and Sances2016; Hauser and Schwarz, Reference Hauser and Schwarz2016, Paolacci et al., Reference Paolacci, Chandler and Ipeirotis2010), or answering honestly (Chandler et al., 2014; Rouse, Reference Rouse2015) or behaving as they normally would.

METHOD

We identified three areas where payment might affect subject behavior that could matter to a researcher: self-selection (who chooses to accept the HIT), engagement (how actively subjects paid attention to and interacted with the study), and performance (how those subjects reacted to what they saw in the study). Since, we can identify no correct form of behavior; we simply look to see if different pay rates produce different between-subject behavior across a range of measures. If pay rates do play an influence, we would expect to see either a linear relationship (where higher rates of pay lead to greater attention and performance), or a threshold effect (where performance shifts when an “acceptable rate” has been reached) on a consistent basis. Thus, we are not seeking a single significant finding, but are looking for emerging patterns of behavioral differences that emerge between pay groups.

We conducted two separate studies – one short and easy, the other long and difficult – in order to view the effects of different pay rates on performance in different styles of social science experiments. The first study was a short survey experiment designed in Qualtrics, involving one randomized image followed by 13 questions.Footnote 1 The second study was programmed in the Dynamic Process Tracing Environment (DPTE) and asked subjects to learn about and vote for political candidates.Footnote 2

If pay rates influence subject recruitment and participation, we anticipate subjects are likely to perform optimally when their compensation is highest (Hus et al., Reference Hus, Schmeiser, Haggerty and Nelson2017; Ye et al., Reference Ye, You and Robert2017). Subjects who feel they are being adequately compensated for their work are more likely to pay attention, to take seriously the task at hand, and to focus on the decisions they are asked to consider. Of course, as the studies progress and subjects spent greater time and effort in participating, their attitudes about “being adequately compensated” may change.

Thus, we further suspect that any differences in subject behavior are more likely to show up later in the study than earlier. Our first study, which took only about 4 min to complete, was unlikely to produce differences in behavior between the beginning and end of the survey. Our second study however, which could take 60 minutes to complete, we believe is more likely to produce effects toward the end of the study as subjects tired of participation and may have begun re-evaluating whether their payment was indeed adequate.

RESULTS

Our results from both studies were roughly identical, in that we found few reportable differences in our measures between the different pay rates.Footnote 3 For brevity, and to save space on reproducing dozens of null results, we only present our second study here, as it permits the more thorough look at Turker behavior. Matching results for the survey experiment can be found in the Online Appendix.

We first examine if our pay rates affected who we recruited to complete our study. We had no a priori assumptions about how pay rates might affect recruitment, so we relied on what we considered to be “conventional” demographic measures that we use in political science.

Table 1Footnote 4 shows that none of our eight categories (percentages of women, African-Americans, Hispanics, Democrats, Independents, or the mean age, political interest, or conservatism of our subjects) return significant results. Further, only one of our categories shows a consistent pattern in the results (a steady increase in Hispanic subjects as pay rates increased). With a relatively small sample size of 364 subjects, it is possible that a larger sample size might produce significant results, but looking at the substantive differences in results, it seems more likely that our demographic measures tended to show random fluctuation between the pay rates, rather than systematic differences in who chose to sign up for the study.

Table 1 Subject Demographics of the DPTE study, by Pay Rate

Our larger concern is for things that we were not able to measure, such as Turker experience. It is possible that more experienced Turkers may gravitate toward higher pay rates, or studies that they feel have a higher pay-to-effort ratio. This is, regrettably, something that we were not able to measure. However, since experimental samples do not tend to seek representative samples on Mechanical Turk, we feel that the risk of any demographic or background differences in who we recruit is that it could then lead to differences in behavior, either through attention to the study or in reaction to the various elements of the study. While we do not find observable demographic differences, we can continue on by examining how people performed within the study.

An advantage of using a DPTE experiment is that we have much greater ability to tease out how subjects performed across a range of measures. We first present the results of our attention checks, and then will move on to discuss engagement with the experiment and candidate evaluation.

Table 2 shows that the vast majority of all of our subjects passed our attention check tests, and there are again no significant differences between our pay rate groups.Footnote 5 There is an apparent pattern of subjects passing at higher rates when paid more however, which suggests that perhaps there may be an effect that our study was not large enough to fully capture. The lowest rates of passing the first two popups in the Primary are found in the $2 pay group (93.8% for both), and while subjects in the higher pay groups all passed the third and fourth popup at a 100% rate, subjects in our minimal $2 pay group passed this at the lowest rates we find in the study, below 90%. While not a significant finding, this suggests that perhaps subjects in this lowest pay group were not paying attention to the extent of the other pay groups.

Table 2 Subject Reaction to Attention Checks, by Pay Rate

If this is the case, however, further evidence should emerge elsewhere. We would expect that attention would get worse as the study carried on. However, it does not. These differences do not appear again in the General Election, when we expected effects to be the greatest. Overall, we find that our subjects generally responded well to our attention checks regardless of what they were being paid.

Beyond merely paying attention to what was presented to them, this study also asked subjects to actively engage with the program, and actively learn about political candidates. This is another area where differential motivation based upon pay rates could influence behavior. Table 3 presents a series of one-way analysis-of-variance tests on measures of active engagement with the experiment. While the previous table measured how much attention subjects paid to the study, this table assesses how actively engaged Turkers were in interacting with the dynamic information boards by selecting information to view. If payments created different incentives to participate, this should be observable through the time subjects spent in the campaign scenarios, the number of items they chose to view, and how much time they devoted to the political aspects of the study relative to the more entertaining current event items.

Table 3 Subject Engagement with the Experiment, by Pay Rate

We find only one statistically significant result, and thus no consistent or clear evidence that pay rates influenced our subject behavior. The lone significant finding we have occurs for our measure of the number of information items subjects chose to open during the Primary Election. While significant, these results show that our highest paid group sought out the most information in the primary, while the second highest group sought out the least. This does not sensibly fit to our theory, and is not replicated along other measures. The lack of a clear pattern within the data again suggests that pay rates did not systematically influence subject performance, even in a long and taxing study.

A final way for us to consider how our subjects participated in the study is to evaluate their final decisions and evaluations of the candidates. It is possible that, while behavioral differences did not emerge, perhaps psychological appraisals of the subject matter were effected by anticipated rewards. We find, again, very little evidence that pay rates mattered. We asked our subjects who they voted for, how confident they were in their vote decision, how difficult that vote choice was, and how much they felt they knew about the candidates, for both the Executive and House race.

The only significant finding we have in Table 4 is for the confidence our subjects had in selecting the House candidate that they truly preferred. Here, we find a significant result and a pattern indicating that lower-paid subjects had greater confidence in their vote choice. This could lead us to assume that our rates of pay influenced how much consideration or psychological investment our subjects had in the study. However, this again appears to be an isolated finding. In all other measures, there are no significant differences or patterns in the data to find that pay rates played a role in how our subjects felt about the candidates or their vote decisions.

Table 4 Subject Evaluation of the Candidates, by Pay Rate

CONCLUSIONS

Our results are quite easy to summarize – pay rates did not seem to matter much to subject performance among Mechanical Turkers, at least not that we observed. While we only discuss our first study here, these results are replicated across another shorter study that collected a much larger sample and is presented in the Online Appendix. In both studies, no systematic patterns emerged that might suggest that pay rates significantly or substantively influenced subject behavior. This does not mean, of course, that pay rates produce no effects, but simply that we, using two very different social science studies, and observing numerous measures of behavior in each, were not able to identify any such effects. We do feel that have observed most, if not all, of the important characteristics of behavior likely to change.

Importantly, we report these results without correcting for multiple hypotheses testing, which would only further reduce the minimal effects we found. In each of our four areas, we analyze we have at least eight different measures, suggesting that by chance alone we should find some significant findings. Indeed, we do. However, these findings show no clear patterns of the influence of pay rates and it is in the absence of patterns that we feel safest in drawing our conclusions. Our clearest path is to conclude that pay rates largely do not influence subject participation and behavior on Mechanical Turk.

This is an important null finding for social scientists using online labor pools. However, we do not intend here to conclude fully that pay rates do not matter. Paying a fair wage for work done does still involve ethical standards (Zechmeister, Reference Zechmeister and Desposato2015). While our discipline as a whole has never established what ethical wages are for subjects, several suggestions both within the Turker community and academic literature have suggested a $6 per hour rate. This still makes crowdsourced samples considerably cheaper than professional alternatives, while also paying a fair rate to the people whose work we depend upon.

SUPPLEMENTARY MATERIAL

To view supplementary material for this article, please visit https://doi.org/10.1017/XPS.2018.7

Footnotes

The data, code, and any additional materials required to replicate all analyses in this article are available at the Journal of Experimental Political Science Dataverse within the Harvard Dataverse Network, at: doi:10.7910/DVN/VCWWGZ.

2 A more thorough description of the study can be found in the online appendix. The HIT we posted and the full study we employed can be viewed online at: https://dpte.polisci.uiowa.edu/dpte/action/player/launch/921/22772?pass=Archived&skip=1

3 The data, syntax, and additional materials required to replicate all analyses in this article are available at the Journal of Experimental Political Science Dataverse within the Harvard Dataverse Network, at: doi:10.7910/DVN/VCWWGZ

4 Pay rates could also influence how fast subjects accept and complete the study, but we found no evidence of this. Every batch we posted completed in approximately the same time, but because of the nature of how AMT posts HITs and reports completions, it is difficult to analyze more precisely. The lower pay rate groups closed slighty slower than the higher pay rate groups, but the substantive difference was minimal and seemed to be caused by subjects accepting the HIT and then waiting to complete it until the time limit was due.

5 Due to a programming glitch, our subjects on the $2 pay day did not see the attention check questions, but they did still view our “pop up” attention checks.

References

REFERENCES

Andersen, David. 2018. “Replication Data for: Subject Performance in Social Science Experiments Using Crowdsources Online Samples.” doi:10.7910/DVN/VCWWGZ, Harvard Dataverse, V1, UNF:6:RQAq0OAZinHNkPjUZVcz5A==.Google Scholar
Berinsky, Adam, Huber, Gregory, and Lenz, Gabriel. 2012. “Evaluating Online Labor Markets for Experimental Research: Amazon.com's Mechanical Turk.” Political Analysis 20: 351368.Google Scholar
Berinsky, Adam, Margolis, Michele, and Sances, Michael. 2016. “Can we turn shirkers into workers?.” Journal of Experimental Social Psychology 66: 2028.Google Scholar
Druckman, James N. and Kam, Cindy D.. 2011. “Students As Experimental Participants: A Defense of the ‘Narrow Data Base’.” In Handbook of Experimental Political Science, eds. Druckman, James N., Green, Donald P., Kuklinski, James H, and Lupia, Arthur. (pp. 4157). New York: Cambridge University Press.Google Scholar
Druckman, James, Green, Donald Kuklinski, James, and Lupia, Arthur 2006. “The Growth and Development of Experimental Research in Political Science.” American Political Science Review 100 (4): 627635.Google Scholar
Finnerty, Ailbhe, Kucherbaev, Pavel, Tranquillini, Stefano, and Convertino, Gregorio. 2013. “Keep it Simple: Reward and Task Design in Crowdsourcing.” Paper presented at CHItaly ‘13, Trento, Italy, September 16–20.Google Scholar
Hauser, David J. and Schwarz, Norbert. 2016. “Attentive Turkers: MTurk Participants Perform Better on Online Attention Checks Than do Subject Pool Participants.” Bahavior Research Methods 48 (1): 400407.Google Scholar
Hitlin, Paul. 2016. ‘Research in the Crowdsourcing Age, a Case Study’ Pew Research Center. July 2016. Available at: http://www.pewinternet.org/2016/07/11/research-in-the-crowdsourcing-age-a-case-study/Google Scholar
Ho, Chien-Ju, Slivkins, Aleksandrs, Suri, Diddharth, and Vaughan, Jennifer Wortman. 2015. Incentivizing high quality crowdwork. Paper presented at the International World Wide Web Conference, Florence, Italy, May 18–22.Google Scholar
Horton, John J. and Chilton, Lydia B.. 2010. “The Labor Economics of Paid Crowdsourcing.” Presented at the 11th ACM conference on electronic commerce (pp. 209–218). Cambridge, Massachusetts: ACM.Google Scholar
Huff, Connor and Tingley, Dustin. 2015. “‘Who are these people?’ Evaluating the demographic characteristics and political preferences of MTurk survey respondents.” Research & Politics 2 (3).Google Scholar
Hus, Joanne W., Schmeiser, Maximilian D., Haggerty, Catherine, and Nelson, Shannon. 2017. “The Effect of Large Monetary Incentives on Survey Completion: Evidence from a Randomized Experiment with the Survey of Consumer Finances.” Public Opinion Quarterly 81 (Fall): 736747.Google Scholar
Ipeirotis, Panagiotis G. 2010. Demographics of Mechanical Turk. NYU Working Paper No. CEDER-10-01. Available at SSRN: https://ssrn.com/abstract=1585030. Accessed March 14, 2018.Google Scholar
Iyengar, Shanto. 2011. “Laboratory Experiments in Political Science.” In Handbook of Experimental Political Science, Eds. Druckman, James, Green, Donald, Kuklinski, James and Lupia, Arthur. New York City: Cambridge University Press.Google Scholar
Kaufman, Nicolas, Schulze, Thimo, and Veit, Daniel. 2011. “More than Fun and Money. Worker Motivation in Crowdsourcing – A Study on Mechanical Turk.” Presented at the during the Proceedings of the Seventeenth Americas Conference on Information Systems. Detroit, Michigan, August 4–7.Google Scholar
Krupnikov, Yanna and Levine, Adam Seth. 2014. “Cross-sample Comparisons and External Validity.” Journal of Experimental Political Science 1: 5980.Google Scholar
Lau, Richard R. 1995. “Information Search During an Election Campaign: Introducing a Process Tracing Methodology for Political Scientists.” In Political Judgment: Structure and Process, Eds. Lodge, M. and McGraw, K. (pp. 179206). Ann Arbor, MI: University of Michigan Press.Google Scholar
Lau, Richard R., Andersen, David J., and Redlawsk, David P.. 2008. “An Exploration of Correct Voting in Recent Presidential Elections.” American Journal of Political Science 52 (2): 395411.Google Scholar
Lau, Richard R. and Redlawsk, David P.. 1997. “Voting Correctly.” American Political Science Review 91 (September): 585599.Google Scholar
Lau, Richard R. and Redlawsk, David P.. 2006. How Voters Decide: Information Processing during Election Campaigns. New York: Cambridge University Press.Google Scholar
Levay, Kevin E., Freese, Jeremy, and Druckman, Jamie. 2016. “The Demographic and Political Composition of Mechanical Turk Samples.” SAGE Open, January-March, 2016, 117.Google Scholar
Mason, Winter and Watts, Duncan. 2009. “Financial Incentives and the “Performance of Crowds.” SIGKDD Explorations 11 (2): 100108.Google Scholar
McCrone, David and Bechhofer, Frank. 2015. Understanding National Identity. Cambridge: Cambridge University Press.Google Scholar
McDermott, Rose. 2002. “Experimental Methods in Political Science.” Annual Review of Political Science 5: 3161.Google Scholar
Morton, Rebecca and Williams, Kenneth. 2010. Experimental Political Science and the Study of Causality: From Nature to the Lab. Cambridge University Press.Google Scholar
Mutz, Dianna. 2011. Population-based Survey Experiments. Princeton, NJ: Princeton University Press.Google Scholar
Paolacci, Gabriele, Chandler, Jesse, and Ipeirotis, Panagiotis. 2010. “Running Experiments on Mechanical Turk.” Judgment and Decision Making, 5 (5).Google Scholar
Rogstadius, Jakob, Kostakos, Vassilis, Kittur, Aniket, Smus, Boris, Laredo, Jim, and Vukovic, Maja. 2011. “An Assessment of Intrinsic and Extrinsic Motivation on Task Performance in Crowdsourcing Markets.” Presented at the Fifth International AAAI Conference on Weblogs and Social Media.Google Scholar
Rouse, Steven V. 2015. “A Reliability Analysis of Mechanical Turk data.” Computers in Human Behavior, 43: 304307.Google Scholar
Schulze, Thimo, Krug, Simone, and Schader, Martin. 2012. “Workers’ Task Choice in Crowdsourcing and Human Computation Markets.” Presented at the thirty third International Conference on Information Systems, held in Orlando, Fl.Google Scholar
Sears, David O. 1986. “College Sophomores in the Laboratory: Influences on a Narrow Data Base on Social Psychology's View on Human Nature.” Journal of Personality and Social Psychology 51 (3): 515530.Google Scholar
Stewart, Neil, Ungemach, Cristoph, Harris, Adam J. L., Bartels, Daniel M., Newll, Ben R., Paolacci, Gabriele, and Chandler, Jesse. 2015. “The Average Laboratory Samples a Population of 7300 Amazon Mechanical Turk Workers.” Judgement and Decision Making 10 (5): 479491.Google Scholar
Ye, Teng, You, Sangseok, and Robert, Lionel P. 2017. “When does more Money Work? Examining the Role of Perceived Fairness in Pay on the Performance of Crowdworkers.” Presented at the Eleventh International AAAI Conference on Web and Social Media.Google Scholar
Zechmeister, Elizabeth. 2015. “Ethics and Research in Political Science: The Responsibilities of the Researcher and the Profession.” In Ethics and Experiments: Problems and Solutions for Social Scientists and Policy Professional, ed. Desposato, Scott. New York, NY: Routledge.Google Scholar
Zizzo, Daniel. 2010. Experimenter Demand Effects in Economic Experiments. Experimental Economics 13 (75).Google Scholar
Figure 0

Table 1 Subject Demographics of the DPTE study, by Pay Rate

Figure 1

Table 2 Subject Reaction to Attention Checks, by Pay Rate

Figure 2

Table 3 Subject Engagement with the Experiment, by Pay Rate

Figure 3

Table 4 Subject Evaluation of the Candidates, by Pay Rate

Supplementary material: File

Andersen and Lau supplementary material

Online Appendix

Download Andersen and Lau supplementary material(File)
File 28.6 KB
Supplementary material: Link

Andersen and Lau Dataset

Link