Crowdsourcing for Usability Testing
While usability evaluation is critical to designing usable websites, traditional usability testing can be both expensive and time consuming. The advent of crowdsourcing platforms such as Amazon Mechanical Turk and CrowdFlower offer an intriguing new avenue for performing remote usability testing with potentially many users, quick turn-around, and significant cost savings. To investigate the potential of such crowdsourced usability testing, we conducted two similar (though not completely parallel) usability studies which evaluated a graduate school’s website: one via a traditional usability lab setting, and the other using crowdsourcing. While we find crowdsourcing exhibits some notable limitations in comparison to the traditional lab environment, its applicability and value for usability testing is clearly evidenced. We discuss both methodological differences for crowdsourced usability testing, as well as empirical contrasts to results from more traditional, face-to-face usability testing.
💡 Research Summary
The paper presents a systematic comparison between traditional laboratory‑based usability testing and crowdsourced usability testing conducted on Amazon Mechanical Turk (MTurk) and CrowdFlower. The authors selected a graduate school website as the test object and designed two parallel studies that, while not perfectly identical, shared the same core tasks: locating departmental information, checking admission procedures, downloading scholarship application forms, and other typical navigation scenarios.
In the laboratory study, eight participants (graduate students and faculty members) were recruited on‑site, worked on a fixed workstation, and were recorded with high‑resolution screen capture, mouse‑tracking, and eye‑tracking equipment. After each task, participants took part in a semi‑structured interview, and the researchers collected quantitative metrics (task completion time, error count, success rate) and qualitative feedback (System Usability Scale (SUS) scores, open‑ended comments).
For the crowdsourced study, the same five tasks were posted as web‑based surveys on MTurk (30 workers) and CrowdFlower (35 workers). Workers received a brief written description, a set of screenshots illustrating the target UI elements, and were asked to perform the tasks on their own devices. The study incorporated “gold questions” (tasks with known correct answers) and “attention checks” to filter low‑quality responses. Compensation was set at an average of $2.5 per worker, and the total cost for both platforms was roughly $150. Data collected consisted of click logs, timestamps, and short textual responses to SUS and open‑ended prompts.
The results reveal stark contrasts across three dimensions: efficiency, data quality, and depth of insight. Crowdsourcing achieved a turnaround time of under two hours for data collection, compared with roughly twelve hours of recruitment, testing, and data processing in the lab. Cost savings were dramatic: $150 versus about $1,200 for the lab study. However, the quality gap was evident. Lab participants completed tasks in an average of 45 seconds with a 12 % error rate, while crowdsourced workers took about 62 seconds and exhibited a 27 % error rate. Errors in the crowdsourced group were heavily concentrated on visual UI elements such as dropdown menus and radio buttons, reflecting uncontrolled variations in screen resolution, browser rendering, and network latency.
Qualitative findings also diverged. In‑person participants provided rich, contextual feedback during interviews, pinpointing specific navigation bottlenecks, confusing terminology, and missing mobile‑responsive features. Crowdsourced participants, constrained by a single text box, offered only brief remarks like “confusing” or “hard to find,” limiting the researchers’ ability to generate actionable design recommendations.
The authors discuss why these differences arise. Laboratory testing offers a controlled environment, direct observation, and the opportunity for iterative probing, which yields high‑fidelity data but at high cost and limited sample size. Crowdsourcing, by contrast, trades control for scale, speed, and diversity, making it well‑suited for early‑stage validation, low‑stakes design checks, and large‑scale quantitative surveys.
Based on their experience, the authors propose a set of practical guidelines for conducting crowdsourced usability tests:
- Task Design – Keep tasks simple, provide annotated screenshots, and break complex flows into discrete steps.
- Quality Assurance – Embed gold questions, attention checks, and automatic outlier detection; consider post‑hoc manual review of ambiguous logs.
- Compensation Strategy – Use performance‑based bonuses (e.g., higher pay for low error rates or fast completion) to motivate careful work.
- Pilot Testing – Run a small pilot on the chosen platform to calibrate instructions, estimate completion time, and refine quality filters.
- Hybrid Workflow – Use crowdsourcing for broad pattern detection and initial problem identification, then follow up with a smaller, controlled lab study to explore the most critical issues in depth.
The paper concludes that crowdsourced usability testing is not a wholesale replacement for traditional methods but a valuable complementary tool, especially for projects constrained by budget or timeline. Future research directions include cross‑platform comparisons (e.g., Prolific, Figure Eight), longitudinal studies tracking user behavior over time, and the integration of machine‑learning models to automatically flag low‑quality responses. By combining the speed and cost advantages of crowdsourcing with the depth and rigor of laboratory testing, designers and researchers can achieve a more efficient and comprehensive usability evaluation pipeline.