skip to main content
10.1145/3306618.3314255acmconferencesArticle/Chapter ViewAbstractPublication PagesaiesConference Proceedingsconference-collections
research-article

Taking Advantage of Multitask Learning for Fair Classification

Authors Info & Claims
Published:27 January 2019Publication History

ABSTRACT

A central goal of algorithmic fairness is to reduce bias in automated decision making. An unavoidable tension exists between accuracy gains obtained by using sensitive information as part of a statistical model, and any commitment to protect these characteristics. Often, due to biases present in the data, using the sensitive information in the functional form of a classifier improves classification accuracy. In this paper we show how it is possible to get the best of both worlds: optimize model accuracy and fairness without explicitly using the sensitive feature in the functional form of the model, thereby treating different individuals equally. Our method is based on two key ideas. On the one hand, we propose to use Multitask Learning (MTL), enhanced with fairness constraints, to jointly learn group specific classifiers that leverage information between sensitive groups. On the other hand, since learning group specific models might not be permitted, we propose to first predict the sensitive features by any learning method and then to use the predicted sensitive feature to train MTL with fairness constraints. This enables us to tackle fairness with a three-pronged approach, that is, by increasing accuracy on each group, enforcing measures of fairness during training, and protecting sensitive information during testing. Experimental results on two real datasets support our proposal, showing substantial improvements in both accuracy and fairness.

References

  1. J. Adebayo and L. Kagal. 2016. Iterative orthogonal feature projection for diagnosing bias in black-box models. In Conference on Fairness, Accountability, and Transparency in Machine Learning (FATML).Google ScholarGoogle Scholar
  2. A. Agarwal, A. Beygelzimer, M. Dud'ik, and J. Langford. 2017. A Reductions Approach to Fair Classification. In FATML.Google ScholarGoogle Scholar
  3. A. Agarwal, A. Beygelzimer, M. Dud'ik, J. Langford, and H. Wallach. 2018. A reductions approach to fair classification. arXiv preprint arXiv:1803.02453 (2018).Google ScholarGoogle Scholar
  4. D. Alabi, N. Immorlica, and A. T. Kalai. 2018. When optimizing nonlinear objectives is no harder than linear objectives. arXiv preprint arXiv:1804.04503 (2018).Google ScholarGoogle Scholar
  5. A. Argyriou, T. Evgeniou, and M. Pontil. 2008. Convex multi-task feature learning. Machine Learning, Vol. 73, 3 (2008), 243--272. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. B. Bakker and T. Heskes. 2003. Task clustering and gating for bayesian multitask learning. Journal of Machine Learning Research, Vol. 4 (2003), 83--99. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. J. Baxter. 2000. A model of inductive bias learning. Journal of artificial intelligence research, Vol. 12 (2000), 149--198. Google ScholarGoogle ScholarCross RefCross Ref
  8. Y. Bechavod and K. Ligett. 2018. Penalizing Unfairness in Binary Classification. arXiv preprint arXiv:1707.00044v3 (2018).Google ScholarGoogle Scholar
  9. R. Berk, H. Heidari, S. Jabbari, M. Joseph, M. Kearns, J. Morgenstern, S. Neel, and A. Roth. 2017. A convex framework for fair regression. arXiv preprint arXiv:1706.02409 (2017).Google ScholarGoogle Scholar
  10. A. Beutel, J. Chen, Z. Zhao, and E. H. Chi. 2017. Data decisions and theoretical implications when adversarially learning fair representations. In FATML.Google ScholarGoogle Scholar
  11. L. Breiman. 2001. Random forests. Machine learning, Vol. 45, 1 (2001), 5--32. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. F. Calmon, D. Wei, B. Vinzamuri, K. Natesan Ramamurthy, and K. R. Varshney. 2017. Optimized Pre-Processing for Discrimination Prevention. In Advances in Neural Information Processing Systems (NIPS). Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. R. Caruana. 1997. Multitask Learning. Machine Learning, Vol. 28, 1 (1997), 41--75. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. A. Chouldechova. 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data, Vol. 5, 2 (2017), 153--163.Google ScholarGoogle Scholar
  15. M. Donini, D. Martinez-Rego, M. Goodson, J. Shawe-Taylor, and M. Pontil. 2016. Distributed variance regularized multitask learning. In International Joint Conference on Neural Networks.Google ScholarGoogle Scholar
  16. M. Donini, L. Oneto, S. Ben-David, J. Shawe-Taylor, and M. Pontil. 2018. Empirical Risk Minimization under Fairness Constraints. In NIPS. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. C. Dwork, N. Immorlica, A. T. Kalai, and M. D. M. Leiserson. 2018. Decoupled Classifiers for Group-Fair and Efficient Machine Learning. In Conference on Fairness, Accountability and Transparency (FAT).Google ScholarGoogle Scholar
  18. T. Evgeniou and M. Pontil. 2004. Regularized multi-task learning. In ACM SIGKDD international conference on Knowledge discovery and data mining. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. M. Feldman, S. A. Friedler, J. Moeller, C. Scheidegger, and S. Venkatasubramanian. 2015. Certifying and removing disparate impact. In International Conference on Knowledge Discovery and Data Mining. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. M. Hardt, E. Price, and N. Srebro. 2016. Equality of opportunity in supervised learning. In NIPS. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. IBM. 2018. User-Manual CPLEX 12.7.1. IBM Software Group.Google ScholarGoogle Scholar
  22. F. Kamiran and T. Calders. 2009. Classifying without discriminating. In International Conference on Computer, Control and Communication.Google ScholarGoogle Scholar
  23. F. Kamiran and T. Calders. 2010. Classification with no discrimination by preferential sampling. In Machine Learning Conference.Google ScholarGoogle Scholar
  24. F. Kamiran and T. Calders. 2012. Data preprocessing techniques for classification without discrimination. Knowledge and Information Systems, Vol. 33, 1 (2012), 1--33. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. T. Kamishima, S. Akaho, and J. Sakuma. 2011. Fairness-aware learning through regularization approach. In International Conference on Data Mining Workshops. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. M. Kearns, S. Neel, A. Roth, and Z. S. Wu. 2017. Preventing Fairness Gerrymandering: Auditing and Learning for Subgroup Fairness. arXiv preprint arXiv:1711.05144 (2017).Google ScholarGoogle Scholar
  27. A. Khosla, T. Zhou, T. Malisiewicz, A. A. Efros, and A. Torralba. 2012. Undoing the damage of dataset bias. In European Conference on Computer Vision. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. A. K. Menon and R. C. Williamson. 2018. The cost of fairness in binary classification. In FAT.Google ScholarGoogle Scholar
  29. D. Pedreshi, S. Ruggieri, and F. Turini. 2008. Discrimination-aware data mining. In ACM SIGKDD international conference on Knowledge discovery and data mining. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. A. Pérez-Suay, V. Laparra, G. Mateo-Garc'ia, J. Mu n oz-Mar'i, L. Gómez-Chova, and G. Camps-Valls. 2017. Fair Kernel Learning. In Machine Learning and Knowledge Discovery in Databases.Google ScholarGoogle Scholar
  31. G. Pleiss, M. Raghavan, F. Wu, J. Kleinberg, and K. Q. Weinberger. 2017. On fairness and calibration. In NIPS. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. S. Shalev-Shwartz and S. Ben-David. 2014. Understanding machine learning: From theory to algorithms .Cambridge University Press. Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. J. Shawe-Taylor and N. Cristianini. 2004. Kernel methods for pattern analysis .Cambridge University Press. Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. A. J. Smola and B. Schölkopf. 2001. Learning with Kernels .MIT Press.Google ScholarGoogle Scholar
  35. B. Woodworth, S. Gunasekar, M. I. Ohannessian, and N. Srebro. 2017. Learning non-discriminatory predictors. In Computational Learning Theory.Google ScholarGoogle Scholar
  36. M. B. Zafar, I. Valera, M. Gomez Rodriguez, and K. P. Gummadi. 2017. Fairness beyond disparate treatment & disparate impact: Learning classification without disparate mistreatment. In International Conference on World Wide Web. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. M. B. Zafar, I. Valera, M. Gomez Rodriguez, and K. P. Gummadi. 2017. Fairness constraints: Mechanisms for fair classification. In International Conference on Artificial Intelligence and Statistics.Google ScholarGoogle Scholar
  38. M. B. Zafar, I. Valera, M. Rodriguez, K. Gummadi, and A. Weller. 2017. From parity to preference-based notions of fairness in classification. In NIPS. Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. R. Zemel, Y. Wu, K. Swersky, T. Pitassi, and C. Dwork. 2013. Learning fair representations. In International Conference on Machine Learning. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Taking Advantage of Multitask Learning for Fair Classification

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          AIES '19: Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society
          January 2019
          577 pages
          ISBN:9781450363242
          DOI:10.1145/3306618

          Copyright © 2019 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 27 January 2019

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article

          Acceptance Rates

          Overall Acceptance Rate61of162submissions,38%

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader