Are online participants still a treasure? Or are they rubbish?

Since the COVID pandemic, the number of online experiments in accounting has been increasing rapidly. Lockdown measures at universities and business schools have pushed experimental researchers out of the lab and toward online participant pools. Field researchers found themselves without partnering firms and institutions and have started fishing from online participant pools to maintain their research pipeline. The increasing number of online experiments has also led to more criticism. In this post, I discuss this criticism and argue that, despite the challenges of recruiting online participants, they can still be a valuable pool, depending on the type and focus of your experiment.

I have experienced the popularity of online experiments as a rollercoaster. I still remember a seminar speaker presenting an online experiment during my doctoral studies years ago. These were the early days of Amazon’s Mechanical Turk (MTurk), and a seminar speaker claimed that the online participants they recruited using MTurk were, to a large degree, real executives participating in MTurk studies during their lunch. This claim was (rightfully) met with much skepticism by the seminar participants. It shows the tone at the time with papers like Farrell, Grenier, and Leiby (2017) defending the online experiment’s honor by suggesting that their quality was just as sound as student samples. During those years, it was challenging to formally critique the use of online participant pools in the experimental accounting community.

Today, little remains of the euphoria from that time, and, as things stand right now, our rollercoaster cart seems to be accelerating fast on its way down. Recent research has revealed that online participants can have multiple accounts, enabling them to participate in the same study multiple times, and there are issues with participant farms, VPN/VPSs, and bots (Ahler et al. 2019, Dennis et al. 2020, Kennedy et al. 2020). The quality of the MTurk participant pool seems to have deteriorated the most, making it challenging to recruit motivated and qualified participants for experimental studies. Ever since these quality issues came to light, an increasing amount of criticism has emerged on the use of online participants in general, even to the extent that choosing online participants in and of itself can be raised as a major concern during discussions and in reviews.

In response to these concerns, many have proposed fixes that should address these problems. Some, for instance, argue that some online platforms and labor markets are better than others. They propose that the only way to get high-quality online participants nowadays is to choose other platforms (e.g., Prolific) and ditch the traditional platforms (e.g., MTurk). To these folks, Prolific solves any problem we might see with recruiting online participants. Others seem more open-minded and believe MTurk can still have value. In addition to switching platforms, other researchers have proposed running studies through parties like Cloud Research or adding extensive pre-screening questions to filter out "poor quality" types (see Bentley 2021). Personally, I am a proponent of using programming to objectively filter participants violating the license agreement of the online platform and the consent form they accepted before taking part in the study. In my recent experiments, I always use Javascript to filter participants relatively effortlessly and without storing personally identifiable data serverside. For more information, please see this post.

Regardless of these fixes, there is little consensus within the community about which fix, or combination of fixes, helps address some or all of the abovementioned challenges with using online participants and why. Thus, they are rather ineffective in slowing our rollercoaster down, making me wonder whether the end of the ride is near or whether we will end up riding our way to the top again once another participant pool enters the market. However, this rollercoaster ride I find myself in does little to no justice to the bigger questions about the use of online participants and the broader trade-offs behind deciding which participant pool to use as an experimental researcher. While I will not go as far as claiming that online participant pools are bulletproof as some of my colleagues did in the early MTurk days, I will use these trade-offs to offer some counterweight to the problems that we are so conveniently using to sell our snake oil fix.

The most important reason favoring online participant recruitment is that it is cheap and fast. However, whether online participants are suitable also depends on the type, focus, and research question of your experimental study. For instance, online participant pools might not be the BEST choice for managerial scenarios and complex audit tasks. It is challenging to ensure online participants take part seriously and are motivates to take part in such studies. Moreover, it is likely that such a heterogenous and noisy participant pool proxies significantly worse for your target population of managers and auditors than, for instance, a student pool or practitioner pool. All the problems associated with recruiting participants online make matters worse. Also, any proposed fix or combination will never really cut it. Your trimmed-down, highly selective sample is unlikely to miraculously reflect the manager or auditor population better than a student or practitioner sample.

However, recruiting online participants may be acceptable for other types of research questions and target populations. For instance, if your experiment features simple effort tasks or everyday decisions that any type of person should relate to, then it may be easier to motivate online participants to take part seriously. Also, your research question is far less restrictive in its target population and recruiting participants online may be perfectly fine. Another example is experiments focusing on retail/unsophisticated investors or the average consumer. These experiments may also benefit from recruiting online participants rather than specific groups, such as students or experts. Some of the fixes may still help improve the quality of your sample. However, many of the remaining control problems and related fixes are just not as critical for this type of research. The quality of your research will never hinge entirely on whether you apply these fixes or not. There is an important difference between trying to recruit "joe the plumber" for a straightforward real effort task and recruiting an auditor for a detailed scenario featuring an audit task.

Economists are often interested in broader and more fundamental research questions about human behavior. Thus, they tend to view online experiments as means to generalize results beyond experimental labs and the (rather restrictive) student samples (Chen et al. 2016, Snowberg and Yariv 2021). Yet, like us they also want to ensure online participants are not just a convenience sample that is particularly characteristic of low quality. One of the solutions they use is throwing quantity at the problem of low quality online participants. Namely, empirical evidence reveals that online samples are primarily more noisier than student samples (Snowberg and Yariv 2021). Noisier dependent variables are not necessarily detrimental to the hypothesis tests of your experiment. It just means you may require a larger sample size to cut through the noise and detect your hypothesized effects.

Conclusion

We should exit the rollercoaster ride. Online participant pools can (still) be a good pond for accounting experimenters to fish from because they offer relatively cheap and a broad range of participants fast. Depending on your research question and target population, these participants may even be right up your experiment’s alley. Suppose you target a heterogenous population or people with no particular prerequisite knowledge or specialization (e.g., consumers, retail investors, etc.). In that case, online participant pools may be preferred over students or professionals burdened by a long-term career.

References

Ahler, D. J., C. E. Roush, & G. Sood. 2019. The micro-task market for lemons: Data quality on Amazon's Mechanical Turk. Political Science Research and Methods, 1-20.
Bentley, J. W. 2021. Improving the statistical power and reliability of research using Amazon Mechanical Turk. Accounting Horizons 35 (4): 45-62.
Chen, D. L., M. Schonger, & C. Wickens. (2016). oTree—An open-source platform for laboratory, online, and field experiments. Journal of Behavioral and Experimental Finance, 9: 88-97.
Dennis, S. A., B. M. Goodson, & C. A. Pearson. 2020. Online worker fraud and evolving threats to the integrity of MTurk data: A discussion of virtual private servers and the limitations of IP-based screening procedures. Behavioral Research in Accounting, 32 (1), 119-134.
Farrell, A. M., J. H. Gernier, and J. Leiby. 2017. Scoundrels or stars? Theory and evidence on the quality of workers in online labor markets. The Accounting Review 92 (1): 93-114.
Kennedy, R., S. Clifford, T. Burleigh, P. D. Waggoner, R. Jewell, & N. J. Winter. 2020. The shape of and solutions to the MTurk quality crisis. Political Science Research and Methods, 8 (4), 614-629.
Snowberg, E., & L. Yariv. (2021). Testing the waters: Behavior across participant pools. American Economic Review, 111 (2): 687-719.

How to reference this online article?

V. F. J. van Pelt (2022, September 24). Are online participants still a treasure? Or are they rubbish? Accounting Experiments, Available at: https://www.accountingexperiments.com/post/online-participants
Avatar
Victor van Pelt
Assistant Professor of Management Accounting

Related