Nowadays assuring that search and recommendation systems are fair and do not apply discrimination among any kind of population has become of paramount importance. Those systems typically rely on machine learning algorithms that solve the classification task. Although the problem of fairness has been widely addressed in binary classification, unfortunately, the fairness of multi-class classification problem needs to be further investigated lacking well-established solutions. For the aforementioned reasons, in this paper, we present the Debiaser for Multiple Variables, a novel approach able to enhance fairness in both binary and multi-class classification problems. The proposed method is compared, under several conditions, with the well-established baseline. We evaluate our method on a heterogeneous data set and prove how it overcomes the established algorithms in the multi-classification setting, while maintaining good performances in binary classification. Finally, we present some limitations and future improvements.

D'Aloisio, G.; Stilo, Giovanni; Di Marco, A.; D'Angelo, A.. (2022). Enhancing Fairness in Classification Tasks with Multiple Variables: A Data- and Model-Agnostic Approach. In Ludovico Boratto, Stefano Faralli, Mirko Marras, Giovanni Stilo (Eds.), Advances in Bias and Fairness in Information Retrieval : Third International Workshop, BIAS 2022, Stavanger, Norway, April 10, 2022, Revised Selected Papers (pp. 117-129). Springer. Isbn: 978-3-031-09315-9. Isbn: 978-3-031-09316-6. Doi: 10.1007/978-3-031-09316-6_11.

Enhancing Fairness in Classification Tasks with Multiple Variables: A Data- and Model-Agnostic Approach

Stilo G.
Methodology
;
2022

Abstract

Nowadays assuring that search and recommendation systems are fair and do not apply discrimination among any kind of population has become of paramount importance. Those systems typically rely on machine learning algorithms that solve the classification task. Although the problem of fairness has been widely addressed in binary classification, unfortunately, the fairness of multi-class classification problem needs to be further investigated lacking well-established solutions. For the aforementioned reasons, in this paper, we present the Debiaser for Multiple Variables, a novel approach able to enhance fairness in both binary and multi-class classification problems. The proposed method is compared, under several conditions, with the well-established baseline. We evaluate our method on a heterogeneous data set and prove how it overcomes the established algorithms in the multi-classification setting, while maintaining good performances in binary classification. Finally, we present some limitations and future improvements.
2022
978-3-031-09315-9
978-3-031-09316-6
Algorithmic Bias; Fairness; Machine Learning
D'Aloisio, G.; Stilo, Giovanni; Di Marco, A.; D'Angelo, A.. (2022). Enhancing Fairness in Classification Tasks with Multiple Variables: A Data- and Model-Agnostic Approach. In Ludovico Boratto, Stefano Faralli, Mirko Marras, Giovanni Stilo (Eds.), Advances in Bias and Fairness in Information Retrieval : Third International Workshop, BIAS 2022, Stavanger, Norway, April 10, 2022, Revised Selected Papers (pp. 117-129). Springer. Isbn: 978-3-031-09315-9. Isbn: 978-3-031-09316-6. Doi: 10.1007/978-3-031-09316-6_11.
File in questo prodotto:
File Dimensione Formato  
_BIAS_2022__Enhancing_Fairness_in_Classification_Tasks_with_Multiple_Variables__a_Data__and_Model_Agnostic_Approach.pdf

Solo gestori archivio

Tipologia: Versione dell'editore
Licenza: Tutti i diritti riservati
Dimensione 380.7 kB
Formato Adobe PDF
380.7 kB Adobe PDF   Visualizza/Apri
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11385/252643
Citazioni
  • Scopus 4
  • ???jsp.display-item.citation.isi??? 2
  • OpenAlex ND
social impact