Reasoning about Political Bias in Content Moderation

Authors

  • Shan Jiang Northeastern University
  • Ronald E. Robertson Northeastern University
  • Christo Wilson Northeastern University

DOI:

https://doi.org/10.1609/aaai.v34i09.7117

Abstract

Content moderation, the AI-human hybrid process of removing (toxic) content from social media to promote community health, has attracted increasing attention from lawmakers due to allegations of political bias. Hitherto, this allegation has been made based on anecdotes rather than logical reasoning and empirical evidence, which motivates us to audit its validity. In this paper, we first introduce two formal criteria to measure bias (i.e., independence and separation) and their contextual meanings in content moderation, and then use YouTube as a lens to investigate if the political leaning of a video plays a role in the moderation decision for its associated comments. Our results show that when justifiable target variables (e.g., hate speech and extremeness) are controlled with propensity scoring, the likelihood of comment moderation is equal across left- and right-leaning videos.

Downloads

Published

2020-04-03

How to Cite

Jiang, S., Robertson, R. E., & Wilson, C. (2020). Reasoning about Political Bias in Content Moderation. Proceedings of the AAAI Conference on Artificial Intelligence, 34(09), 13669-13672. https://doi.org/10.1609/aaai.v34i09.7117