Skip to main content
Erschienen in: Data Mining and Knowledge Discovery 5/2023

14.02.2023

Enforcing fairness using ensemble of diverse Pareto-optimal models

verfasst von: Vitória Guardieiro, Marcos M. Raimundo, Jorge Poco

Erschienen in: Data Mining and Knowledge Discovery | Ausgabe 5/2023

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

One of the main challenges of machine learning is to ensure that its applications do not generate or propagate unfair discrimination based on sensitive characteristics such as gender, race, and ethnicity. Research in this area typically limits models to a level of discrimination quantified by an equity metric (usually the “benefit” discrepancy between privileged and non-privileged groups). However, when models reduce bias, they may also reduce their performance (e.g., accuracy, F1 score). Therefore, we have to optimize contradictory metrics (performance and fairness) at the same time. This problem is well characterized as a multi-objective optimization (MOO) problem. In this study, we use MOO methods to minimize the difference between groups, maximize the benefits for each group, and preserve performance. We search for the best trade-off models in binary classification problems and aggregate them using ensemble filtering and voting procedures. The aggregation of models with different levels of benefits for each group improves robustness regarding performance and fairness. We compared our approach with other known methodologies, using logistic regression as a benchmark for comparison. The proposed methods obtained interesting results: (i) multi-objective training found models that are similar to or better than the adversarial methods and are more diverse in terms of fairness and accuracy metrics, (ii) multi-objective selection was able to improve the balance between fairness and accuracy compared to selection with a single metric, and (iii) the final predictor found models with higher fairness without sacrificing much accuracy.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
For two possible parameters \(\varvec{\theta }_i\) and \(\varvec{\theta }_j\), it is said that \(\varvec{\theta }_i\) weakly dominates \(\varvec{\theta }_j\), noted as \(G(\varvec{\theta }_i) \preceq G(\varvec{\theta }_j)\), if \(g_k(\varvec{\theta }_i) \le g_k(\varvec{\theta }_j), \forall k \in {1,\dots , m}\).
 
Literatur
Zurück zum Zitat Abdi H (2010) Coefficient of variation. Encycl Res Design 1:169–171 Abdi H (2010) Coefficient of variation. Encycl Res Design 1:169–171
Zurück zum Zitat Abebe SA, Lucchese C, Orlando S (2022) Eifffel: enforcing fairness in forests by flipping leaves. In: Proceedings of the 37th ACM/SIGAPP symposium on applied computing, pp. 429–436 Abebe SA, Lucchese C, Orlando S (2022) Eifffel: enforcing fairness in forests by flipping leaves. In: Proceedings of the 37th ACM/SIGAPP symposium on applied computing, pp. 429–436
Zurück zum Zitat Agarwal A, Beygelzimer A, Dudik M, Langford J, Wallach H (2018) A reductions approach to fair classification. In: Dy J, Krause A (eds.) Proceedings of the 35th international conference on machine learning, Proceedings of machine learning research, vol. 80, pp. 60–69. PMLR. http://proceedings.mlr.press/v80/agarwal18a.html Agarwal A, Beygelzimer A, Dudik M, Langford J, Wallach H (2018) A reductions approach to fair classification. In: Dy J, Krause A (eds.) Proceedings of the 35th international conference on machine learning, Proceedings of machine learning research, vol. 80, pp. 60–69. PMLR. http://​proceedings.​mlr.​press/​v80/​agarwal18a.​html
Zurück zum Zitat Bellamy RK, Dey K, Hind M, Hoffman SC, Houde S, Kannan K, Lohia P, Martino J, Mehta S, Mojsilović A et al (2019) Ai fairness 360: an extensible toolkit for detecting and mitigating algorithmic bias. IBM J Res Dev 63(4/5):1–4CrossRef Bellamy RK, Dey K, Hind M, Hoffman SC, Houde S, Kannan K, Lohia P, Martino J, Mehta S, Mojsilović A et al (2019) Ai fairness 360: an extensible toolkit for detecting and mitigating algorithmic bias. IBM J Res Dev 63(4/5):1–4CrossRef
Zurück zum Zitat Berk R, Heidari H, Jabbari S, Joseph M, Kearns M, Morgenstern J, Neel S, Roth A (2017) A convex framework for fair regression. arXiv preprint arXiv:1706.02409 Berk R, Heidari H, Jabbari S, Joseph M, Kearns M, Morgenstern J, Neel S, Roth A (2017) A convex framework for fair regression. arXiv preprint arXiv:​1706.​02409
Zurück zum Zitat Bhargava V, Couceiro M, Napoli A (2020) Limeout: an ensemble approach to improve process fairness. In: Joint European conference on machine learning and knowledge discovery in databases, pp. 475–491. Springer Bhargava V, Couceiro M, Napoli A (2020) Limeout: an ensemble approach to improve process fairness. In: Joint European conference on machine learning and knowledge discovery in databases, pp. 475–491. Springer
Zurück zum Zitat Bhaskaruni D, Hu H, Lan C (2019) Improving prediction fairness via model ensemble. In: 2019 IEEE 31st International conference on tools with artificial intelligence (ICTAI), pp. 1810–1814. IEEE Bhaskaruni D, Hu H, Lan C (2019) Improving prediction fairness via model ensemble. In: 2019 IEEE 31st International conference on tools with artificial intelligence (ICTAI), pp. 1810–1814. IEEE
Zurück zum Zitat Binns R (2020) On the apparent conflict between individual and group fairness. In: Proceedings of the 2020 conference on fairness, accountability, and transparency, pp. 514–524 Binns R (2020) On the apparent conflict between individual and group fairness. In: Proceedings of the 2020 conference on fairness, accountability, and transparency, pp. 514–524
Zurück zum Zitat Calders T, Kamiran F, Pechenizkiy M (2009) Building classifiers with independency constraints. In: 2009 IEEE International conference on data mining workshops, pp. 13–18. IEEE Calders T, Kamiran F, Pechenizkiy M (2009) Building classifiers with independency constraints. In: 2009 IEEE International conference on data mining workshops, pp. 13–18. IEEE
Zurück zum Zitat Chen Z, Zhang J, Sarro F, Harman M (2022) Maat: a novel ensemble approach to addressing fairness and performance bugs for machine learning software. In: The ACM joint european software engineering conference and symposium on the foundations of software engineering (ESEC/FSE) Chen Z, Zhang J, Sarro F, Harman M (2022) Maat: a novel ensemble approach to addressing fairness and performance bugs for machine learning software. In: The ACM joint european software engineering conference and symposium on the foundations of software engineering (ESEC/FSE)
Zurück zum Zitat Cohon JL (2004) Multiobjective programming and planning, vol. 140. Courier Corporation Cohon JL (2004) Multiobjective programming and planning, vol. 140. Courier Corporation
Zurück zum Zitat Cohon JL, Church RL, Sheer DP (1979) Generating multiobjective trade-offs: an algorithm for bicriterion problems. Water Resour Res 15(5):1001–1010CrossRef Cohon JL, Church RL, Sheer DP (1979) Generating multiobjective trade-offs: an algorithm for bicriterion problems. Water Resour Res 15(5):1001–1010CrossRef
Zurück zum Zitat Corbett-Davies S, Goel S (2018) The measure and mismeasure of fairness: a critical review of fair machine learning. arXiv preprint arXiv:1808.00023 Corbett-Davies S, Goel S (2018) The measure and mismeasure of fairness: a critical review of fair machine learning. arXiv preprint arXiv:​1808.​00023
Zurück zum Zitat Cruz AF, Saleiro P, Belém C, Soares C, Bizarro P (2020) A bandit-based algorithm for fairness-aware hyperparameter optimization. arXiv preprint arXiv:2010.03665 Cruz AF, Saleiro P, Belém C, Soares C, Bizarro P (2020) A bandit-based algorithm for fairness-aware hyperparameter optimization. arXiv preprint arXiv:​2010.​03665
Zurück zum Zitat d’Alessandro B, O’Neil C, LaGatta T (2017) Conscientious classification: A data scientist’s guide to discrimination-aware classification. Big Data 5(2):120–134CrossRef d’Alessandro B, O’Neil C, LaGatta T (2017) Conscientious classification: A data scientist’s guide to discrimination-aware classification. Big Data 5(2):120–134CrossRef
Zurück zum Zitat Dieterich W, Mendoza C, Brennan T (2016) Compas risk scales: demonstrating accuracy equity and predictive parity. Northpoint Inc 7(74), 1 Dieterich W, Mendoza C, Brennan T (2016) Compas risk scales: demonstrating accuracy equity and predictive parity. Northpoint Inc 7(74), 1
Zurück zum Zitat Dressel J, Farid H (2018) The accuracy, fairness, and limits of predicting recidivism. Sci Adv 4(1):eaao5580CrossRef Dressel J, Farid H (2018) The accuracy, fairness, and limits of predicting recidivism. Sci Adv 4(1):eaao5580CrossRef
Zurück zum Zitat Dutta S, Wei D, Yueksel H, Chen PY, Liu S, Varshney K (2020) Is there a trade-off between fairness and accuracy? a perspective using mismatched hypothesis testing. In: International conference on machine learning, pp. 2803–2813. PMLR Dutta S, Wei D, Yueksel H, Chen PY, Liu S, Varshney K (2020) Is there a trade-off between fairness and accuracy? a perspective using mismatched hypothesis testing. In: International conference on machine learning, pp. 2803–2813. PMLR
Zurück zum Zitat Dwork C, Hardt M, Pitassi T, Reingold O, Zemel R (2012) Fairness through awareness. In: Proceedings of the 3rd innovations in theoretical computer science conference, pp. 214–226 Dwork C, Hardt M, Pitassi T, Reingold O, Zemel R (2012) Fairness through awareness. In: Proceedings of the 3rd innovations in theoretical computer science conference, pp. 214–226
Zurück zum Zitat Grgic-Hlaca N, Zafar MB, Gummadi KP, Weller A (2016) The case for process fairness in learning: feature selection for fair decision making. In: NIPS symposium on machine learning and the Law, vol. 1, p. 2 Grgic-Hlaca N, Zafar MB, Gummadi KP, Weller A (2016) The case for process fairness in learning: feature selection for fair decision making. In: NIPS symposium on machine learning and the Law, vol. 1, p. 2
Zurück zum Zitat Grgic-Hlaca N, Zafar MB, Gummadi KP, Weller A (2017) On fairness, diversity and randomness in algorithmic decision making. CoRR Grgic-Hlaca N, Zafar MB, Gummadi KP, Weller A (2017) On fairness, diversity and randomness in algorithmic decision making. CoRR
Zurück zum Zitat Hardt M, Price E, Srebro N (2016) Equality of opportunity in supervised learning. Adv Neural Inform Process Syst 29 Hardt M, Price E, Srebro N (2016) Equality of opportunity in supervised learning. Adv Neural Inform Process Syst 29
Zurück zum Zitat Howard A, Borenstein J (2018) The ugly truth about ourselves and our robot creations: the problem of bias and social inequity. Sci Eng Ethics 24(5):1521–1536CrossRef Howard A, Borenstein J (2018) The ugly truth about ourselves and our robot creations: the problem of bias and social inequity. Sci Eng Ethics 24(5):1521–1536CrossRef
Zurück zum Zitat Iosifidis V, Ntoutsi E (2019) Adafair: cumulative fairness adaptive boosting. In: Proceedings of the 28th ACM international conference on information and knowledge management, pp. 781–790 Iosifidis V, Ntoutsi E (2019) Adafair: cumulative fairness adaptive boosting. In: Proceedings of the 28th ACM international conference on information and knowledge management, pp. 781–790
Zurück zum Zitat Kamiran F, Calders T (2012) Data preprocessing techniques for classification without discrimination. Knowl Inf Syst 33(1):1–33CrossRef Kamiran F, Calders T (2012) Data preprocessing techniques for classification without discrimination. Knowl Inf Syst 33(1):1–33CrossRef
Zurück zum Zitat Kamishima T, Akaho S, Asoh H, Sakuma J (2012) Fairness-aware classifier with prejudice remover regularizer. In: Joint European conference on machine learning and knowledge discovery in databases, pp. 35–50. Springer Kamishima T, Akaho S, Asoh H, Sakuma J (2012) Fairness-aware classifier with prejudice remover regularizer. In: Joint European conference on machine learning and knowledge discovery in databases, pp. 35–50. Springer
Zurück zum Zitat Kamishima T, Akaho S, Sakuma J (2011) Fairness-aware learning through regularization approach. In: 2011 IEEE 11th international conference on data mining workshops, pp. 643–650. IEEE Kamishima T, Akaho S, Sakuma J (2011) Fairness-aware learning through regularization approach. In: 2011 IEEE 11th international conference on data mining workshops, pp. 643–650. IEEE
Zurück zum Zitat Kearns M, Roth A (2019) The ethical algorithm: the science of socially aware algorithm design. Oxford University Press Kearns M, Roth A (2019) The ethical algorithm: the science of socially aware algorithm design. Oxford University Press
Zurück zum Zitat Kenfack PJ, Khan AM, Kazmi SA, Hussain R, Oracevic A, Khattak AM (2021) Impact of model ensemble on the fairness of classifiers in machine learning. In: 2021 International conference on applied artificial intelligence (ICAPAI), pp. 1–6. IEEE Kenfack PJ, Khan AM, Kazmi SA, Hussain R, Oracevic A, Khattak AM (2021) Impact of model ensemble on the fairness of classifiers in machine learning. In: 2021 International conference on applied artificial intelligence (ICAPAI), pp. 1–6. IEEE
Zurück zum Zitat Kusner MJ, Loftus J, Russell C, Silva R (2017) Counterfactual fairness. Adv Neural Inform Process Syst 30 Kusner MJ, Loftus J, Russell C, Silva R (2017) Counterfactual fairness. Adv Neural Inform Process Syst 30
Zurück zum Zitat Liu S, Vicente LN (2022) Accuracy and fairness trade-offs in machine learning: a stochastic multi-objective approach. Comput Manag Sci pp. 1–25 Liu S, Vicente LN (2022) Accuracy and fairness trade-offs in machine learning: a stochastic multi-objective approach. Comput Manag Sci pp. 1–25
Zurück zum Zitat Martinez N, Bertran M, Sapiro G (2020) Minimax pareto fairness: A multi objective perspective. In: H.D. III, A. Singh (eds.) Proceedings of the 37th international conference on machine learning, Proceedings of machine learning research, vol. 119, pp. 6755–6764. PMLR. http://proceedings.mlr.press/v119/martinez20a.html Martinez N, Bertran M, Sapiro G (2020) Minimax pareto fairness: A multi objective perspective. In: H.D. III, A. Singh (eds.) Proceedings of the 37th international conference on machine learning, Proceedings of machine learning research, vol. 119, pp. 6755–6764. PMLR. http://​proceedings.​mlr.​press/​v119/​martinez20a.​html
Zurück zum Zitat Mehrabi N, Morstatter F, Saxena N, Lerman K, Galstyan A (2021) A survey on bias and fairness in machine learning. ACM Comput Surv (CSUR) 54(6):1–35CrossRef Mehrabi N, Morstatter F, Saxena N, Lerman K, Galstyan A (2021) A survey on bias and fairness in machine learning. ACM Comput Surv (CSUR) 54(6):1–35CrossRef
Zurück zum Zitat Miettinen K (2012) Nonlinear multiobjective optimization, vol. 12. Springer Science & Business Media Miettinen K (2012) Nonlinear multiobjective optimization, vol. 12. Springer Science & Business Media
Zurück zum Zitat Osoba OA, Welser IV W (2017) An intelligence in our image: the risks of bias and errors in artificial intelligence. Rand Corporation Osoba OA, Welser IV W (2017) An intelligence in our image: the risks of bias and errors in artificial intelligence. Rand Corporation
Zurück zum Zitat Padh K, Antognini D, Lejal-Glaude E, Faltings B. Musat C (2021) Addressing fairness in classification with a model-agnostic multi-objective algorithm. In: Uncertainty in artificial intelligence, pp. 600–609. PMLR Padh K, Antognini D, Lejal-Glaude E, Faltings B. Musat C (2021) Addressing fairness in classification with a model-agnostic multi-objective algorithm. In: Uncertainty in artificial intelligence, pp. 600–609. PMLR
Zurück zum Zitat Raimundo MM, Von Zuben FJ (2020) Multi-criteria analysis involving pareto-optimal misclassification tradeoffs on imbalanced datasets. In: 2020 international joint conference on neural networks (IJCNN), pp. 1–8. IEEE Raimundo MM, Von Zuben FJ (2020) Multi-criteria analysis involving pareto-optimal misclassification tradeoffs on imbalanced datasets. In: 2020 international joint conference on neural networks (IJCNN), pp. 1–8. IEEE
Zurück zum Zitat Raimundo MM, Drumond TF, Marques ACR, Lyra C, Rocha A, Von Zuben FJ (2021) Exploring multiobjective training in multiclass classification. Neurocomputing 435:307–320CrossRef Raimundo MM, Drumond TF, Marques ACR, Lyra C, Rocha A, Von Zuben FJ (2021) Exploring multiobjective training in multiclass classification. Neurocomputing 435:307–320CrossRef
Zurück zum Zitat Savic D (2002) Single-objective vs. multiobjective optimisation for integrated decision support. Proc First Bienn Meet Int Environ Model Softw Soc 1:7–12 Savic D (2002) Single-objective vs. multiobjective optimisation for integrated decision support. Proc First Bienn Meet Int Environ Model Softw Soc 1:7–12
Zurück zum Zitat Speicher T, Heidari H, Grgic-Hlaca N, Gummadi KP, Singla A, Weller A, Zafar MB (2018) A unified approach to quantifying algorithmic unfairness: Measuring individual & group unfairness via inequality indices. In: Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining, KDD ’18, p. 2239-2248. Association for computing machinery Speicher T, Heidari H, Grgic-Hlaca N, Gummadi KP, Singla A, Weller A, Zafar MB (2018) A unified approach to quantifying algorithmic unfairness: Measuring individual & group unfairness via inequality indices. In: Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining, KDD ’18, p. 2239-2248. Association for computing machinery
Zurück zum Zitat Wadsworth C, Vera F, Piech C (2018) Achieving fairness through adversarial learning: an application to recidivism prediction Wadsworth C, Vera F, Piech C (2018) Achieving fairness through adversarial learning: an application to recidivism prediction
Zurück zum Zitat Zafar MB, Valera I, Rodriguez MG, Gummadi KP, Weller A (2017) From parity to preference-based notions of fairness in classification Zafar MB, Valera I, Rodriguez MG, Gummadi KP, Weller A (2017) From parity to preference-based notions of fairness in classification
Zurück zum Zitat Zafar MB, Valera I, Rodriguez M, Gummadi K, Weller A (2017) From parity to preference-based notions of fairness in classification. Adv Neural Inform Process Syst 30 Zafar MB, Valera I, Rodriguez M, Gummadi K, Weller A (2017) From parity to preference-based notions of fairness in classification. Adv Neural Inform Process Syst 30
Zurück zum Zitat Zafar MB, Valera I, Rogriguez MG, Gummadi KP (2017) Fairness constraints: mechanisms for fair classification. In: Artificial intelligence and statistics, pp. 962–970. PMLR Zafar MB, Valera I, Rogriguez MG, Gummadi KP (2017) Fairness constraints: mechanisms for fair classification. In: Artificial intelligence and statistics, pp. 962–970. PMLR
Zurück zum Zitat Zemel R, Wu Y, Swersky K, Pitassi T, Dwork C (2013) Learning fair representations. In: International conference on machine learning, pp. 325–333. PMLR Zemel R, Wu Y, Swersky K, Pitassi T, Dwork C (2013) Learning fair representations. In: International conference on machine learning, pp. 325–333. PMLR
Zurück zum Zitat Zhang W, Bifet A, Zhang X, Weiss JC, Nejdl W (2021) Farf: a fair and adaptive random forests classifier. In: Pacific-Asia conference on knowledge discovery and data mining, pp. 245–256. Springer Zhang W, Bifet A, Zhang X, Weiss JC, Nejdl W (2021) Farf: a fair and adaptive random forests classifier. In: Pacific-Asia conference on knowledge discovery and data mining, pp. 245–256. Springer
Zurück zum Zitat Zhang Q, Liu J, Zhang Z, Wen J, Mao B, Yao X (2021) Fairer machine learning through multi-objective evolutionary learning. In: International conference on artificial neural networks, pp. 111–123. Springer Zhang Q, Liu J, Zhang Z, Wen J, Mao B, Yao X (2021) Fairer machine learning through multi-objective evolutionary learning. In: International conference on artificial neural networks, pp. 111–123. Springer
Zurück zum Zitat Zhang Q, Liu J, Zhang Z, Wen J, Mao B, Yao X (2022) Mitigating unfairness via evolutionary multi-objective ensemble learning. In: IEEE transactions on evolutionary computation Zhang Q, Liu J, Zhang Z, Wen J, Mao B, Yao X (2022) Mitigating unfairness via evolutionary multi-objective ensemble learning. In: IEEE transactions on evolutionary computation
Zurück zum Zitat Zhang W, Weiss JC (2021) Fair decision-making under uncertainty. In: 2021 IEEE international conference on data mining (ICDM), pp. 886–895. IEEE Zhang W, Weiss JC (2021) Fair decision-making under uncertainty. In: 2021 IEEE international conference on data mining (ICDM), pp. 886–895. IEEE
Zurück zum Zitat Zhang W, Weiss JC (2022) Longitudinal fairness with censorship. In: Proceedings of the AAAI conference on artificial intelligence, vol. 36, pp. 12235–12243 Zhang W, Weiss JC (2022) Longitudinal fairness with censorship. In: Proceedings of the AAAI conference on artificial intelligence, vol. 36, pp. 12235–12243
Zurück zum Zitat Zhao H, Gordon G (2019) Inherent tradeoffs in learning fair representations. Adv Neural Inform Process Syst 32 Zhao H, Gordon G (2019) Inherent tradeoffs in learning fair representations. Adv Neural Inform Process Syst 32
Zurück zum Zitat Zliobaite I (2015) On the relation between accuracy and fairness in binary classification Zliobaite I (2015) On the relation between accuracy and fairness in binary classification
Metadaten
Titel
Enforcing fairness using ensemble of diverse Pareto-optimal models
verfasst von
Vitória Guardieiro
Marcos M. Raimundo
Jorge Poco
Publikationsdatum
14.02.2023
Verlag
Springer US
Erschienen in
Data Mining and Knowledge Discovery / Ausgabe 5/2023
Print ISSN: 1384-5810
Elektronische ISSN: 1573-756X
DOI
https://doi.org/10.1007/s10618-023-00922-y

Weitere Artikel der Ausgabe 5/2023

Data Mining and Knowledge Discovery 5/2023 Zur Ausgabe

Premium Partner