TY - GEN
T1 - Tracking and improving information in the service of fairness
AU - Garg, Sumegha
AU - Kim, Michael P.
AU - Reingold, Omer
N1 - Publisher Copyright:
© 2019 Association for Computing Machinery.
PY - 2019/6/17
Y1 - 2019/6/17
N2 - As algorithmic prediction systems have become widespread, fears that these systems may inadvertently discriminate against members of underrepresented populations have grown. With the goal of understanding fundamental principles that underpin the growing number of approaches to mitigating algorithmic discrimination, we investigate the role of information in fair prediction. A common strategy for decision-making uses a predictor to assign individuals a risk score; then, individuals are selected or rejected on the basis of this score. In this work, we study a formal framework for measuring the information content of predictors. Central to the framework is the notion of a refinement; intuitively, a refinement of a predictor z increases the overall informativeness of the predictions without losing the information already contained in z. We show that increasing information content through refinements improves the downstream selection rules across a wide range of fairness measures (e.g. true positive rates, false positive rates, selection rates). In turn, refinements provide a simple but effective tool for reducing disparity in treatment and impact without sacrificing the utility of the predictions. Our results suggest that in many applications, the perceived "cost of fairness" results from an information disparity across populations, and thus, may be avoided with improved information.
AB - As algorithmic prediction systems have become widespread, fears that these systems may inadvertently discriminate against members of underrepresented populations have grown. With the goal of understanding fundamental principles that underpin the growing number of approaches to mitigating algorithmic discrimination, we investigate the role of information in fair prediction. A common strategy for decision-making uses a predictor to assign individuals a risk score; then, individuals are selected or rejected on the basis of this score. In this work, we study a formal framework for measuring the information content of predictors. Central to the framework is the notion of a refinement; intuitively, a refinement of a predictor z increases the overall informativeness of the predictions without losing the information already contained in z. We show that increasing information content through refinements improves the downstream selection rules across a wide range of fairness measures (e.g. true positive rates, false positive rates, selection rates). In turn, refinements provide a simple but effective tool for reducing disparity in treatment and impact without sacrificing the utility of the predictions. Our results suggest that in many applications, the perceived "cost of fairness" results from an information disparity across populations, and thus, may be avoided with improved information.
KW - Algorithmic fairness
KW - Prediction
UR - http://www.scopus.com/inward/record.url?scp=85069042091&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85069042091&partnerID=8YFLogxK
U2 - 10.1145/3328526.3329624
DO - 10.1145/3328526.3329624
M3 - Conference contribution
AN - SCOPUS:85069042091
T3 - ACM EC 2019 - Proceedings of the 2019 ACM Conference on Economics and Computation
SP - 809
EP - 824
BT - ACM EC 2019 - Proceedings of the 2019 ACM Conference on Economics and Computation
PB - Association for Computing Machinery, Inc
T2 - 20th ACM Conference on Economics and Computation, EC 2019
Y2 - 24 June 2019 through 28 June 2019
ER -