Fairness in Machine Learning: Against False Positive Rate Equality as a Measure of Fairness

As machine learning informs increasingly consequential decisions, different metrics have been proposed for measuring algorithmic bias or unfairness. Two popular “fairness measures” are calibration and equality of false positive rate. Each measure seems intuitively important, but notably, it is usual...

全面介绍

Saved in:  
书目详细资料
主要作者: Long, Robert (Author)
格式: 电子 文件
语言:English
Check availability: HBZ Gateway
Journals Online & Print:
载入...
Fernleihe:Fernleihe für die Fachinformationsdienste
出版: Brill 2022
In: Journal of moral philosophy
Year: 2022, 卷: 19, 发布: 1, Pages: 49-78
Further subjects:B statistical discrimination
B 公平
B algorithmic bias
在线阅读: Presumably Free Access
Volltext (lizenzpflichtig)
Volltext (lizenzpflichtig)
实物特征
总结:As machine learning informs increasingly consequential decisions, different metrics have been proposed for measuring algorithmic bias or unfairness. Two popular “fairness measures” are calibration and equality of false positive rate. Each measure seems intuitively important, but notably, it is usually impossible to satisfy both measures. For this reason, a large literature in machine learning speaks of a “fairness tradeoff” between these two measures. This framing assumes that both measures are, in fact, capturing something important. To date, philosophers have seldom examined this crucial assumption, and examined to what extent each measure actually tracks a normatively important property. This makes this inevitable statistical conflict – between calibration and false positive rate equality – an important topic for ethics. In this paper, I give an ethical framework for thinking about these measures and argue that, contrary to initial appearances, false positive rate equality is in fact morally irrelevant and does not measure fairness.
ISSN:1745-5243
Contains:Enthalten in: Journal of moral philosophy
Persistent identifiers:DOI: 10.1163/17455243-20213439