How Crowd Worker Factors Influence Subjective Annotations: A Study of Tagging Misogynistic Hate Speech in Tweets

Authors

  • Danula Hettiachchi RMIT University
  • Indigo Holcombe-James RMIT University
  • Stephanie Livingstone RMIT University
  • Anjalee de Silva The University of Melbourne
  • Matthew Lease University of Texas at Austin
  • Flora D. Salim University of New South Wales
  • Mark Sanderson RMIT University

DOI:

https://doi.org/10.1609/hcomp.v11i1.27546

Keywords:

Crowdsourcing, Subjective Annotation Tasks, Annotator Bias, Hate Speech, Content Moderation

Abstract

Crowdsourced annotation is vital to both collecting labelled data to train and test automated content moderation systems and to support human-in-the-loop review of system decisions. However, annotation tasks such as judging hate speech are subjective and thus highly sensitive to biases stemming from annotator beliefs, characteristics and demographics. We conduct two crowdsourcing studies on Mechanical Turk to examine annotator bias in labelling sexist and misogynistic hate speech. Results from 109 annotators show that annotator political inclination, moral integrity, personality traits, and sexist attitudes significantly impact annotation accuracy and the tendency to tag content as hate speech. In addition, semi-structured interviews with nine crowd workers provide further insights regarding the influence of subjectivity on annotations. In exploring how workers interpret a task — shaped by complex negotiations between platform structures, task instructions, subjective motivations, and external contextual factors — we see annotations not only impacted by worker factors but also simultaneously shaped by the structures under which they labour.

Downloads

Published

2023-11-03

How to Cite

Hettiachchi, D., Holcombe-James, I., Livingstone, S., de Silva, A., Lease, M., Salim, F. D., & Sanderson, M. (2023). How Crowd Worker Factors Influence Subjective Annotations: A Study of Tagging Misogynistic Hate Speech in Tweets. Proceedings of the AAAI Conference on Human Computation and Crowdsourcing, 11(1), 38-50. https://doi.org/10.1609/hcomp.v11i1.27546