English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT
 
 
DownloadE-Mail
  Discrimination in Algorithmic Decision Making: From Principles to Measures and Mechanisms

Zafar, M. B. (2019). Discrimination in Algorithmic Decision Making: From Principles to Measures and Mechanisms. PhD Thesis, Universität des Saarlandes, Saarbrücken. doi:10.22028/D291-27726.

Item is

Files

show Files

Locators

show
hide
Description:
-
OA-Status:
Green

Creators

show
hide
 Creators:
Zafar, Muhammad Bilal1, Author           
Gummadi, Krishna P.1, Advisor           
Gomez Rodriguez, Manuel2, Referee           
Goel, Shared 3, Referee
Affiliations:
1Group K. Gummadi, Max Planck Institute for Software Systems, Max Planck Society, ou_2105291              
2Group M. Gomez Rodriguez, Max Planck Institute for Software Systems, Max Planck Society, ou_2105290              
3External Organizations, ou_persistent22              

Content

show
hide
Free keywords: -
 Abstract: The rise of algorithmic decision making in a variety of applications has also raised concerns about its potential for discrimination against certain social groups. However, incorporating nondiscrimination goals into the design of algorithmic decision making systems (or, classifiers) has proven to be quite challenging. These challenges arise mainly due to the computational complexities involved in the process, and the inadequacy of existing measures to computationally capture discrimination in various situations. The goal of this thesis is to tackle these problems. First, with the aim of incorporating existing measures of discrimination (namely, disparate treatment and disparate impact) into the design of well-known classifiers, we introduce a mechanism of decision boundary covariance, that can be included in the formulation of any convex boundary-based classifier in the form of convex constraints. Second, we propose alternative measures of discrimination. Our first proposed measure, disparate mistreatment, is useful in situations when unbiased ground truth training data is available. The other two measures, preferred treatment and preferred impact, are useful in situations when feature and class distributions of different social groups are significantly different, and can additionally help reduce the cost of nondiscrimination (as compared to the existing measures). We also design mechanisms to incorporate these new measures into the design of convex boundary-based classifiers.

Details

show
hide
Language(s): eng - English
 Dates: 2019-02-0420192019
 Publication Status: Issued
 Pages: 111 p.
 Publishing info: Saarbrücken : Universität des Saarlandes
 Table of Contents: -
 Rev. Type: -
 Identifiers: BibTex Citekey: zafarphd2019
DOI: 10.22028/D291-27726
URN: nbn:de:bsz:291--ds-277263
 Degree: PhD

Event

show

Legal Case

show

Project information

show

Source

show