Please use this identifier to cite or link to this item: https://doi.org/10.21256/zhaw-25909
Full metadata record
DC FieldValueLanguage
dc.contributor.authorBaumann, Joachim-
dc.contributor.authorHannák, Anikó-
dc.contributor.authorHeitz, Christoph-
dc.date.accessioned2022-11-03T15:39:04Z-
dc.date.available2022-11-03T15:39:04Z-
dc.date.issued2022-
dc.identifier.isbn978-1-4503-9352-2de_CH
dc.identifier.urihttps://digitalcollection.zhaw.ch/handle/11475/25909-
dc.description.abstractBinary decision making classifiers are not fair by default. Fairness requirements are an additional element to the decision making rationale, which is typically driven by maximizing some utility function. In that sense, algorithmic fairness can be formulated as a constrained optimization problem. This paper contributes to the discussion on how to implement fairness, focusing on the fairness concepts of positive predictive value (PPV) parity, false omission rate (FOR) parity, and sufficiency (which combines the former two). We show that group-specific threshold rules are optimal for PPV parity and FOR parity, similar to well-known results for other group fairness criteria. However, depending on the underlying population distributions and the utility function, we find that sometimes an upper-bound threshold rule for one group is optimal: utility maximization under PPV parity (or FOR parity) might thus lead to selecting the individuals with the smallest utility for one group, instead of selecting the most promising individuals. This result is counter-intuitive and in contrast to the analogous solutions for statistical parity and equality of opportunity. We also provide a solution for the optimal decision rules satisfying the fairness constraint sufficiency. We show that more complex decision rules are required and that this leads to within-group unfairness for all but one of the groups. We illustrate our findings based on simulated and real data.de_CH
dc.language.isoende_CH
dc.publisherAssociation for Computing Machineryde_CH
dc.rightsLicence according to publishing contractde_CH
dc.subjectComputing methodologyde_CH
dc.subjectMachine learningde_CH
dc.subjectApplied computingde_CH
dc.subjectDecision analysisde_CH
dc.subjectGroup fairness metricsde_CH
dc.subjectAlgorithmic fairnessde_CH
dc.subjectPrediction-based decision makingde_CH
dc.subject.ddc005: Computerprogrammierung, Programme und Datende_CH
dc.subject.ddc658.403: Entscheidungsfindung, Informationsmanagementde_CH
dc.titleEnforcing group fairness in algorithmic decision making : utility maximization under sufficiencyde_CH
dc.typeKonferenz: Paperde_CH
dcterms.typeTextde_CH
zhaw.departementSchool of Engineeringde_CH
zhaw.organisationalunitInstitut für Datenanalyse und Prozessdesign (IDP)de_CH
zhaw.publisher.placeNew Yorkde_CH
dc.identifier.doi10.1145/3531146.3534645de_CH
dc.identifier.doi10.21256/zhaw-25909-
zhaw.conference.details5th ACM Conference on Fairness, Accountability, and Transparency (FAccT), Seoul, Republic of Korea, 21-24 June 2022de_CH
zhaw.funding.euNode_CH
zhaw.originated.zhawYesde_CH
zhaw.pages.end2326de_CH
zhaw.pages.start2315de_CH
zhaw.publication.statusacceptedVersionde_CH
zhaw.publication.reviewPeer review (Publikation)de_CH
zhaw.title.proceedingsProceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparencyde_CH
zhaw.funding.snf187473de_CH
zhaw.funding.zhawAlgorithmic Fairness in data-based decision making: Combining ethics and technologyde_CH
zhaw.author.additionalNode_CH
zhaw.display.portraitYesde_CH
Appears in collections:Publikationen School of Engineering

Files in This Item:
File Description SizeFormat 
2022_Baumann-Hannak-Heitz_Enforcing-group-fairness-algorithmic-decision-making_ACM.pdfAccepted Version2.07 MBAdobe PDFThumbnail
View/Open
Show simple item record
Baumann, J., Hannák, A., & Heitz, C. (2022). Enforcing group fairness in algorithmic decision making : utility maximization under sufficiency [Conference paper]. Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, 2315–2326. https://doi.org/10.1145/3531146.3534645
Baumann, J., Hannák, A. and Heitz, C. (2022) ‘Enforcing group fairness in algorithmic decision making : utility maximization under sufficiency’, in Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency. New York: Association for Computing Machinery, pp. 2315–2326. Available at: https://doi.org/10.1145/3531146.3534645.
J. Baumann, A. Hannák, and C. Heitz, “Enforcing group fairness in algorithmic decision making : utility maximization under sufficiency,” in Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, 2022, pp. 2315–2326. doi: 10.1145/3531146.3534645.
BAUMANN, Joachim, Anikó HANNÁK und Christoph HEITZ, 2022. Enforcing group fairness in algorithmic decision making : utility maximization under sufficiency. In: Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency. Conference paper. New York: Association for Computing Machinery. 2022. S. 2315–2326. ISBN 978-1-4503-9352-2
Baumann, Joachim, Anikó Hannák, and Christoph Heitz. 2022. “Enforcing Group Fairness in Algorithmic Decision Making : Utility Maximization under Sufficiency.” Conference paper. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, 2315–26. New York: Association for Computing Machinery. https://doi.org/10.1145/3531146.3534645.
Baumann, Joachim, et al. “Enforcing Group Fairness in Algorithmic Decision Making : Utility Maximization under Sufficiency.” Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, Association for Computing Machinery, 2022, pp. 2315–26, https://doi.org/10.1145/3531146.3534645.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.