skip to main content
10.1145/3340531.3412705acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
research-article

LiFT: A Scalable Framework for Measuring Fairness in ML Applications

Published:19 October 2020Publication History

ABSTRACT

Many internet applications are powered by machine learned models, which are usually trained on labeled datasets obtained through user feedback signals or human judgments. Since societal biases may be present in the generation of such datasets, it is possible for the trained models to be biased, thereby resulting in potential discrimination and harms for disadvantaged groups. Motivated by the need to understand and address algorithmic bias in web-scale ML systems and the limitations of existing fairness toolkits, we present the LinkedIn Fairness Toolkit (LiFT), a framework for scalable computation of fairness metrics as part of large ML systems. We highlight the key requirements in deployed settings, and present the design of our fairness measurement system. We discuss the challenges encountered in incorporating fairness tools in practice and the lessons learned during deployment at LinkedIn. Finally, we provide open problems based on practical experience.

Skip Supplemental Material Section

Supplemental Material

3340531.3412705.mp4

mp4

93.8 MB

References

  1. M. Abadi et al. TensorFlow: A system for large-scale machine learning. In OSDI, 2016.Google ScholarGoogle Scholar
  2. J. A. Adebayo. FairML: ToolBox for diagnosing bias in predictive modeling. PhD thesis, Massachusetts Institute of Technology, 2016.Google ScholarGoogle Scholar
  3. A. Agarwal, A. Beygelzimer, M. Dudík, J. Langford, and H.Wallach. A reductions approach to fair classification. In ICML, 2018.Google ScholarGoogle Scholar
  4. M. Armbrust et al. Spark SQL: Relational data processing in Spark. In SIGMOD, 2015.Google ScholarGoogle Scholar
  5. N. Bantilan. Themis-ML: A fairness-aware machine learning interface for end-toend discrimination discovery and mitigation. Journal of Technology in Human Services, 36(1), 2018.Google ScholarGoogle ScholarCross RefCross Ref
  6. S. Barocas and M. Hardt. Fairness in machine learning. In NeurIPS Tutorial, 2017.Google ScholarGoogle Scholar
  7. R. K. Bellamy et al. AI Fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias. arXiv preprint arXiv:1810.01943, 2018.Google ScholarGoogle Scholar
  8. A. Beutel, J. Chen, T. Doshi, H. Qian, A.Woodruff, C. Luu, P. Kreitmann, J. Bischof, and E. H. Chi. Putting fairness principles into practice: Challenges, metrics, and improvements. In AIES, 2019.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. A. Chouldechova. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data, 5(2), 2017.Google ScholarGoogle Scholar
  10. K. Crawford. The trouble with bias. In NeurIPS Invited Talk, 2017.Google ScholarGoogle Scholar
  11. C. DiCiccio, S. Vasudevan, K. Basu, K. Kenthapadi, and D. Agarwal. Evaluating fairness using permutation tests. In KDD, 2020.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. S. A. Friedler, C. Scheidegger, S. Venkatasubramanian, S. Choudhary, E. P. Hamilton, and D. Roth. A comparative study of fairness-enhancing interventions in machine learning. In FAT*, 2019.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. S. Galhotra, Y. Brun, and A. Meliou. Fairness testing: Testing software for discrimination. In ESEC/FSE, 2017.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. S. C. Geyik, S. Ambler, and K. Kenthapadi. Fairness-aware ranking in search & recommendation systems with application to LinkedIn talent search. In KDD, 2019.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. P. Good. Permutation Tests: A Practical Guide to Resampling Methods for Testing Hypotheses. Springer series in statistics. Springer, 2000.Google ScholarGoogle ScholarCross RefCross Ref
  16. B. Green and Y. Chen. Disparate interactions: An algorithm-in-the-loop analysis of fairness in risk assessments. In FAT*, 2019.Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. N. Grgic-Hlaca, E. M. Redmiles, K. P. Gummadi, and A. Weller. Human perceptions of fairness in algorithmic decision making: A case study of criminal risk prediction. In WWW, 2018.Google ScholarGoogle Scholar
  18. M. Hardt, E. Price, and N. Srebro. Equality of opportunity in supervised learning. In NIPS, 2016.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. K. Holstein, J. Wortman Vaughan, H. Daumé III, M. Dudik, and H. Wallach. Improving fairness in machine learning systems: What do industry practitioners need? In CHI, 2019.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. F. Kamiran and T. Calders. Data preprocessing techniques for classification without discrimination. Knowledge and Information Systems, 33(1):1--33, 2012.Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. T. Kamishima, S. Akaho, H. Asoh, and J. Sakuma. Fairness-aware classifier with prejudice remover regularizer. In ECML PKDD, 2012.Google ScholarGoogle ScholarCross RefCross Ref
  22. R. Kohavi. Scaling up the accuracy of Naive-Bayes classifiers: A decision-tree hybrid. In KDD, 1996.Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. M. Mitchell, S.Wu, A. Zaldivar, P. Barnes, L. Vasserman, B. Hutchinson, E. Spitzer, I. D. Raji, and T. Gebru. Model cards for model reporting. In FAT*, 2019.Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. M. Ojala and G. C. Garriga. Permutation tests for studying classifier performance. Journal of Machine Learning Research, 11, 2010.Google ScholarGoogle Scholar
  25. F. Pedregosa et al. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12, 2011.Google ScholarGoogle Scholar
  26. pymetrics. audit-AI, 2018. https://github.com/pymetrics/audit-ai.Google ScholarGoogle Scholar
  27. P. Saleiro, B. Kuester, A. Stevens, A. Anisfeld, L. Hinkson, J. London, and R. Ghani. Aequitas: A bias and fairness audit toolkit. arXiv preprint arXiv:1811.05577, 2018.Google ScholarGoogle Scholar
  28. T. Speicher, H. Heidari, N. Grgic-Hlaca, K. P. Gummadi, A. Singla, A. Weller, and M. B. Zafar. A unified approach to quantifying algorithmic unfairness: Measuring individual & group unfairness via inequality indices. In KDD, 2018.Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. M. Srivastava, H. Heidari, and A. Krause. Mathematical notions vs. human perception of fairness: A descriptive approach to fairness for machine learning. In KDD, 2019.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. A. Swami, S. Vasudevan, and J. Huyn. Data sentinel: A declarative production scale data validation platform. In ICDE, 2020.Google ScholarGoogle ScholarCross RefCross Ref
  31. F. Tramer, V. Atlidakis, R. Geambasu, D. Hsu, J.-P. Hubaux, M. Humbert, A. Juels, and H. Lin. FairTest: Discovering unwarranted associations in data-driven applications. In Euro S&P, 2017.Google ScholarGoogle ScholarCross RefCross Ref
  32. M. Zaharia et al. Apache spark: a unified engine for big data processing. Communications of the ACM, 59(11), 2016.Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. M. Zehlike, C. Castillo, F. Bonchi, S. Hajian, and M. Megahed. Fairness Measures: Datasets and software for detecting algorithmic discrimination, 2017. http://fairness-measures.org/.Google ScholarGoogle Scholar

Index Terms

  1. LiFT: A Scalable Framework for Measuring Fairness in ML Applications

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Conferences
            CIKM '20: Proceedings of the 29th ACM International Conference on Information & Knowledge Management
            October 2020
            3619 pages
            ISBN:9781450368599
            DOI:10.1145/3340531

            Copyright © 2020 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 19 October 2020

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article

            Acceptance Rates

            Overall Acceptance Rate1,861of8,427submissions,22%

            Upcoming Conference

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader