Abstract

Organizations often employ data-driven models to inform decisions that can have a significant impact on people's lives (e.g., university admissions, hiring). In order to protect people's privacy and prevent discrimination, these decision-makers may choose to delete or avoid collecting social category data, like sex and race. In this article, we argue that such censoring can exacerbate discrimination by making biases more difficult to detect. We begin by detailing how computerized decisions can lead to biases in the absence of social category data and in some contexts, may even sustain biases that arise by random chance. We then show how proactively using social category data can help illuminate and combat discriminatory practices, using cases from education and employment that lead to strategies for detecting and preventing discrimination. We conclude that discrimination can occur in any sociotechnical system in which someone decides to use an algorithmic process to inform decision-making, and we offer a set of broader implications for researchers and policymakers.

Original languageEnglish (US)
Pages (from-to)78-115
Number of pages38
JournalJournal of Information Policy
Volume8
DOIs
StatePublished - 2018

Keywords

  • Algorithmic discrimination
  • Omitted variables
  • Personal data
  • Signaling
  • Statistical discrimination

ASJC Scopus subject areas

  • Sociology and Political Science
  • Public Administration

Fingerprint

Dive into the research topics of 'Challenges, solutions, and policy implications'. Together they form a unique fingerprint.

Cite this