Explaining contributions of features towards unfairness in classifiers: A novel threshold-dependent Shapley value-based approach
dc.contributor.author | Pelegrina G.D. | |
dc.contributor.author | Siraj S. | |
dc.contributor.author | Duarte L.T. | |
dc.contributor.author | Grabisch M. | |
dc.date.accessioned | 2024-11-01T06:11:44Z | |
dc.date.available | 2024-11-01T06:11:44Z | |
dc.date.issued | 2024 | |
dc.description.abstract | © 2024 Elsevier LtdA number of approaches has been proposed to investigate and mitigate unfairness in machine learning algorithms. However, as the definition and understanding of fairness may vary in different situations, the study of ethical disparities remains an open area of research. Besides the importance of analyzing ethical disparities, explainability in machine learning is also a relevant issue in Trustworthy Artificial Intelligence. Usually, both fairness and explainability analysis are based on a fixed decision threshold, which differentiates the positive cases from the negative ones according to the predicted probabilities. In this paper, we investigate how changes in this threshold can impact the fairness of predictions between protected and other groups and how features contribute towards such a measure. We propose a novel Shapley value-based approach as a tool to investigate how changes in the threshold values change the contribution of each feature towards unfairness. This gives us an ability to evaluate how fairness measures vary for different threshold values and which features have the higher (or lower) impact on creating ethical disparities. We demonstrate this using three different case studies that are carefully chosen to highlight different unfairness scenarios and features contributions. We also applied our proposal as a feature selection strategy, which contributed to decrease unfair results substantially. | |
dc.description.volume | 138 | |
dc.identifier.doi | 10.1016/j.engappai.2024.109427 | |
dc.identifier.issn | None | |
dc.identifier.uri | https://dspace.mackenzie.br/handle/10899/39676 | |
dc.relation.ispartof | Engineering Applications of Artificial Intelligence | |
dc.rights | Acesso Restrito | |
dc.subject.otherlanguage | Fairness | |
dc.subject.otherlanguage | Feature contribution | |
dc.subject.otherlanguage | Interpretable machine learning | |
dc.subject.otherlanguage | Shapley value | |
dc.title | Explaining contributions of features towards unfairness in classifiers: A novel threshold-dependent Shapley value-based approach | |
dc.type | Artigo | |
local.scopus.citations | 0 | |
local.scopus.eid | 2-s2.0-85205447975 | |
local.scopus.subject | Decision threshold | |
local.scopus.subject | Fairness | |
local.scopus.subject | Fairness measures | |
local.scopus.subject | Feature contribution | |
local.scopus.subject | Interpretable machine learning | |
local.scopus.subject | Machine learning algorithms | |
local.scopus.subject | Machine-learning | |
local.scopus.subject | Shapley value | |
local.scopus.subject | Threshold-value | |
local.scopus.subject | Value-based approach | |
local.scopus.updated | 2024-12-01 | |
local.scopus.url | https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85205447975&origin=inward |