Your Behavior Signals Your Reliability: Modeling Crowd Behavioral Traces to Ensure Quality Relevance Annotations

Authors

  • Tanya Goyal University of Texas at Austin
  • Tyler McDonnell University of Texas at Austin
  • Mucahid Kutlu Qatar University
  • Tamer Elsayed Qatar University
  • Matthew Lease University of Texas at Austin

DOI:

https://doi.org/10.1609/hcomp.v6i1.13331

Abstract

While peer-agreement and gold checks are well-established methods for ensuring quality in crowdsourced data collection, we explore a relatively new direction for quality control: estimating work quality directly from workers’ behavioral traces collected during annotation. We propose three behavior-based models to predict label correctness and worker accuracy, then further apply model predictions to label aggregation and optimization of label collection. As part of this work, we collect and share a new Mechanical Turk dataset of behavioral signals judging the relevance of search results. Results show that behavioral data can be effectively used to predict work quality, which could be especially useful with single labeling or in a cold start scenario in which individuals’ prior work history is unavailable. We further show improvement in label aggregation and reducing labeling cost while ensuring data quality.

Downloads

Published

2018-06-15

How to Cite

Goyal, T., McDonnell, T., Kutlu, M., Elsayed, T., & Lease, M. (2018). Your Behavior Signals Your Reliability: Modeling Crowd Behavioral Traces to Ensure Quality Relevance Annotations. Proceedings of the AAAI Conference on Human Computation and Crowdsourcing, 6(1), 41-49. https://doi.org/10.1609/hcomp.v6i1.13331