skip to primary navigation skip to content

CBSU bibliography search


To request a reprint of a CBSU publication, please click here to send us an email (reprints may not be available for all publications)

Realistic precision & accuracy of online experiment platforms, web-browsers, and devices
Authors:
ANWYL-IRVINE, A., DALMAIJER, E.S., Hodges, N., Evershed, J.K.
Reference:
Behavior Research Methods, 02 Nov 2020, 53(4):1407-1425
Year of publication:
2020
CBU number:
8571
Abstract:
Due to increasing ease-of-use and ability to quickly collect large samples, online behavioral research is currently booming. With this popularity, it is important that researchers are aware of who online participants are, and what devices and software they use to access experiments. While it is somewhat obvious that these factors can impact data quality, it remains unclear how big this problem is. To understand how these characteristics impact experiment presentation and data quality, we performed a battery of automated tests on a number of realistic setups. We investigated how different web-building platforms (Gorilla v.20190828, jsPsych v6.0.5, Lab.js v19.1.0, and psychoJS/PsychoPy3 v3.1.5), browsers (Chrome, Edge, Firefox, and Safari), and operating systems (macOS and Windows 10) impact display time across 30 different frame durations for each software combination. We then employed a robot actuator in realistic setups to measure response recording across aforementioned platforms, and between different keyboard types (desktop and integrated laptop). Finally, we analysed data from over 200 000 participants on their demographics, technology, and software to provide context to our findings. We found that modern web-platforms provide a reasonable accuracy and precision for display duration and manual response time, and that no single platform stands out as the best in all features and conditions. In addition, our online participant analysis shows what equipment they are likely to use. Introduction Conducting behavioural research online has vastly increased in the last few years. For instance, the number of papers tracked by Web of Science with the keywords ‘MTurk’ or ‘Mechanical Turk’ (Amazon’s popular platform for accessing online participants or workers, available since 2005) was 642 in 2018, over a five-fold increase over five years from 121 publications in 2013 (Figure 1). While scientist do not exclusively use MTurk for psychological experiments, it is indicative of a trend. For example, Bohannon (2016) reported that published MTurk studies in social science increased from 61 in 2011 to 1200 in 2015 – an almost 20-fold increase. A unique problem with internet-based testing is its reliance on participants’ hardware and software. Researchers who are used to lab-based testing will be intimately familiar with their computer, stimulus software, and hardware, for response collection. At the very least, they can be sure that all participants are tested using the very same system. For online testing, the exact opposite is true: Participants use their own computer (desktop, laptop, tablet, or even phone), with their own operating system, and access experiments through a variety of web browsers. In addition to participant degrees of freedom, researchers can choose between various options to generate experiments. These vary from programming libraries (e.g. jsPsych) to graphical experiment builders (e.g. Gorilla Experiment Builder), and come with their own idiosyncrasies with respect to timing, presentation of visual and auditory stimuli, and response collection. Data for this paper can be found at: https://osf.io/rn9zd
URL:
Data for this project is available at: https://osf.io/rn9zd


genesis();