Skip to main content

2021 | OriginalPaper | Buchkapitel

Fairness in Cardiac MR Image Analysis: An Investigation of Bias Due to Data Imbalance in Deep Learning Based Segmentation

verfasst von : Esther Puyol-Antón, Bram Ruijsink, Stefan K. Piechnik, Stefan Neubauer, Steffen E. Petersen, Reza Razavi, Andrew P. King

Erschienen in: Medical Image Computing and Computer Assisted Intervention – MICCAI 2021

Verlag: Springer International Publishing

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

The subject of ‘fairness’ in artificial intelligence (AI) refers to assessing AI algorithms for potential bias based on demographic characteristics such as race and gender, and the development of algorithms to address this bias. Most applications to date have been in computer vision, although some work in healthcare has started to emerge. The use of deep learning (DL) in cardiac MR segmentation has led to impressive results in recent years, and such techniques are starting to be translated into clinical practice. However, no work has yet investigated the fairness of such models. In this work, we perform such an analysis for racial/gender groups, focusing on the problem of training data imbalance, using a nnU-Net model trained and evaluated on cine short axis cardiac MR data from the UK Biobank dataset, consisting of 5,903 subjects from 6 different racial groups. We find statistically significant differences in Dice performance between different racial groups. To reduce the racial bias, we investigated three strategies: (1) stratified batch sampling, in which batch sampling is stratified to ensure balance between racial groups; (2) fair meta-learning for segmentation, in which a DL classifier is trained to classify race and jointly optimized with the segmentation model; and (3) protected group models, in which a different segmentation model is trained for each racial group. We also compared the results to the scenario where we have a perfectly balanced database. To assess fairness we used the standard deviation (SD) and skewed error ratio (SER) of the average Dice values. Our results demonstrate that the racial bias results from the use of imbalanced training data, and that all proposed bias mitigation strategies improved fairness, with the best SD and SER resulting from the use of protected group models.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
In fair AI, the protected attribute(s) are the ones for which fairness needs to be ensured, e.g. gender or race. A set of samples with the same value(s) for the protected attribute(s) are known as a protected group.
 
Literatur
1.
Zurück zum Zitat Bernard, O., et al.: Deep learning techniques for automatic MRI cardiac multi-structures segmentation and diagnosis: is the problem solved? IEEE Trans. Med. Imaging 37(11), 2514–2525 (2018)CrossRef Bernard, O., et al.: Deep learning techniques for automatic MRI cardiac multi-structures segmentation and diagnosis: is the problem solved? IEEE Trans. Med. Imaging 37(11), 2514–2525 (2018)CrossRef
2.
Zurück zum Zitat Buolamwini, J., Gebru, T.: Gender shades: intersectional accuracy disparities in commercial gender classification. In: Conference on Fairness, Accountability And Transparency, pp. 77–91. PMLR (2018) Buolamwini, J., Gebru, T.: Gender shades: intersectional accuracy disparities in commercial gender classification. In: Conference on Fairness, Accountability And Transparency, pp. 77–91. PMLR (2018)
4.
Zurück zum Zitat Dwork, C., et al.: Fairness through awareness. In: Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, pp. 214–226 (2012) Dwork, C., et al.: Fairness through awareness. In: Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, pp. 214–226 (2012)
5.
6.
Zurück zum Zitat Huang, G., et al.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700–4708 (2017) Huang, G., et al.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700–4708 (2017)
7.
8.
Zurück zum Zitat Isensee, F., et al.: nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 18(2), 203–211 (2021)CrossRef Isensee, F., et al.: nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 18(2), 203–211 (2021)CrossRef
10.
Zurück zum Zitat Kamiran, F., Karim, A., Zhang, X.: Decision theory for discrimination-aware classification. In: 2012 IEEE 12th International Conference on Data Mining, pp. 924–929. IEEE (2012) Kamiran, F., Karim, A., Zhang, X.: Decision theory for discrimination-aware classification. In: 2012 IEEE 12th International Conference on Data Mining, pp. 924–929. IEEE (2012)
11.
Zurück zum Zitat Kishi, S., et al.: Race-ethnic and sex differences in left ventricular structure and function: the coronary artery risk development in young adults (CARDIA) study. J. Am. Heart Assoc. 4(3), e001264 (2015)CrossRef Kishi, S., et al.: Race-ethnic and sex differences in left ventricular structure and function: the coronary artery risk development in young adults (CARDIA) study. J. Am. Heart Assoc. 4(3), e001264 (2015)CrossRef
13.
Zurück zum Zitat Mody, P., et al.: Most important articles on cardiovascular disease among racial and ethnic minorities. Circ.: Cardiovasc. Qual. Outcomes 5(4), e33–e41 (2012) Mody, P., et al.: Most important articles on cardiovascular disease among racial and ethnic minorities. Circ.: Cardiovasc. Qual. Outcomes 5(4), e33–e41 (2012)
14.
Zurück zum Zitat Ngxande, M., Tapamo, J.R., Burke, M.: Bias remediation in driver drowsiness detection systems using generative adversarial networks. IEEE Access 8, 55592–55601 (2020)CrossRef Ngxande, M., Tapamo, J.R., Burke, M.: Bias remediation in driver drowsiness detection systems using generative adversarial networks. IEEE Access 8, 55592–55601 (2020)CrossRef
15.
Zurück zum Zitat Obermeyer, Z., et al.: Dissecting racial bias in an algorithm used to manage the health of populations. Science 366(6464), 447–453 (2019)CrossRef Obermeyer, Z., et al.: Dissecting racial bias in an algorithm used to manage the health of populations. Science 366(6464), 447–453 (2019)CrossRef
16.
Zurück zum Zitat Petersen, S.E., et al.: UK Biobank’s cardiovascular magnetic resonance protocol. J. Cardiovasc. Magn. Reson. 18(1), 1–7 (2015)CrossRef Petersen, S.E., et al.: UK Biobank’s cardiovascular magnetic resonance protocol. J. Cardiovasc. Magn. Reson. 18(1), 1–7 (2015)CrossRef
18.
Zurück zum Zitat Raisi-Estabragh, Z., et al.: Variation of cardiac magnetic resonance radiomics features by age and sex in healthy participants from the UK Biobank. Eur. Heart J. 41(Supplement\(\_\)2), ehaa946-0197 (2020) Raisi-Estabragh, Z., et al.: Variation of cardiac magnetic resonance radiomics features by age and sex in healthy participants from the UK Biobank. Eur. Heart J. 41(Supplement\(\_\)2), ehaa946-0197 (2020)
19.
Zurück zum Zitat Ruijsink, B., Puyol-Antón, E., et al.: Fully automated, quality-controlled cardiac analysis from CMR: validation and large-scale application to characterize cardiac function. Cardiovasc. Imaging 13(3), 684–695 (2020) Ruijsink, B., Puyol-Antón, E., et al.: Fully automated, quality-controlled cardiac analysis from CMR: validation and large-scale application to characterize cardiac function. Cardiovasc. Imaging 13(3), 684–695 (2020)
20.
Zurück zum Zitat Wang, M., Deng, W.: Mitigating bias in face recognition using skewness-aware reinforcement learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9322–9331 (2020) Wang, M., Deng, W.: Mitigating bias in face recognition using skewness-aware reinforcement learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9322–9331 (2020)
21.
Zurück zum Zitat Wang, Z., Qinami, K., et al.: Towards fairness in visual recognition: effective strategies for bias mitigation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8919–8928 (2020) Wang, Z., Qinami, K., et al.: Towards fairness in visual recognition: effective strategies for bias mitigation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8919–8928 (2020)
22.
Zurück zum Zitat Wilder, B., et al.: Clinical trial of an AI-augmented intervention for HIV prevention in youth experiencing homelessness. arXiv preprint arXiv:2009.09559 (2020) Wilder, B., et al.: Clinical trial of an AI-augmented intervention for HIV prevention in youth experiencing homelessness. arXiv preprint arXiv:​2009.​09559 (2020)
25.
Zurück zum Zitat Zhang, B.H., Lemoine, B., Mitchell, M.: Mitigating unwanted biases with adversarial learning. In: Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, pp. 335–340 (2018) Zhang, B.H., Lemoine, B., Mitchell, M.: Mitigating unwanted biases with adversarial learning. In: Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, pp. 335–340 (2018)
Metadaten
Titel
Fairness in Cardiac MR Image Analysis: An Investigation of Bias Due to Data Imbalance in Deep Learning Based Segmentation
verfasst von
Esther Puyol-Antón
Bram Ruijsink
Stefan K. Piechnik
Stefan Neubauer
Steffen E. Petersen
Reza Razavi
Andrew P. King
Copyright-Jahr
2021
DOI
https://doi.org/10.1007/978-3-030-87199-4_39