Skip to main content

06.01.2024 | Research

Enhancing the fairness of offensive memes detection models by mitigating unintended political bias

verfasst von: Gitanjali Kumari, Anubhav Sinha, Asif Ekbal, Arindam Chatterjee, Vinutha B N

Erschienen in: Journal of Intelligent Information Systems

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

This paper tackles the critical challenge of detecting and mitigating unintended political bias in offensive meme detection. Political memes are a powerful tool that can be used to influence public opinion and disrupt voters’ mindsets. However, current visual-linguistic models for offensive meme detection exhibit unintended bias and struggle to accurately classify non-offensive and offensive memes. This can harm the fairness of the democratic process either by targeting minority groups or promoting harmful political ideologies. With Hindi being the fifth most spoken language globally and having a significant number of native speakers, it is essential to detect and remove Hindi-based offensive memes to foster a fair and equitable democratic process. To address these concerns, we propose three debiasing techniques to mitigate the overrepresentation of majority group perspectives while addressing the suppression of minority opinions in political discourse. To support our approach, we curate a comprehensive dataset called Pol_Off_Meme, designed especially for the Hindi language. Empirical analysis of this dataset demonstrates the efficacy of our proposed debiasing techniques in reducing political bias in internet memes, promoting a fair and equitable democratic environment. Our debiased model, named \(DRTIM^{Adv}_{Att}\), exhibited superior performance compared to the CLIP-based baseline model. It achieved a significant improvement of +9.72% in the F1-score while reducing the False Positive Rate Difference (FPRD) by -16% and the False Negative Rate Difference (FNRD) by -14.01%. Our efforts strive to cultivate a more informed and inclusive political discourse, ensuring that all opinions, irrespective of their majority or minority status, receive adequate attention and representation.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
Zurück zum Zitat Aksenov, D., Bourgonje, P., Zaczynska, K., et al. (2021). Fine-grained classification of political bias in German news: A data set and initial experiments. In: Proceedings of the 5th workshop on online abuse and Harms (WOAH 2021). Association for Computational Linguistics, Online, (pp. 121–131). https://doi.org/10.18653/v1/2021.woah-1.13 Aksenov, D., Bourgonje, P., Zaczynska, K., et al. (2021). Fine-grained classification of political bias in German news: A data set and initial experiments. In: Proceedings of the 5th workshop on online abuse and Harms (WOAH 2021). Association for Computational Linguistics, Online, (pp. 121–131). https://​doi.​org/​10.​18653/​v1/​2021.​woah-1.​13
Zurück zum Zitat Arango, A., Perez-Martin, J., & Labrada, A. (2022). HateU at SemEval-2022 task 5: Multimedia automatic misogyny identification. In: Proceedings of the 16th international workshop on semantic evaluation (SemEval-2022). Association for Computational Linguistics, Seattle, United States, (pp. 581–584). https://doi.org/10.18653/v1/2022.semeval-1.80 Arango, A., Perez-Martin, J., & Labrada, A. (2022). HateU at SemEval-2022 task 5: Multimedia automatic misogyny identification. In: Proceedings of the 16th international workshop on semantic evaluation (SemEval-2022). Association for Computational Linguistics, Seattle, United States, (pp. 581–584). https://​doi.​org/​10.​18653/​v1/​2022.​semeval-1.​80
Zurück zum Zitat Badjatiya, P., Gupta, M., & Varma, V. (2020). Stereotypical bias removal for hate speech detection task using knowledge-based generalizations. arXiv:2001.05495 Badjatiya, P., Gupta, M., & Varma, V. (2020). Stereotypical bias removal for hate speech detection task using knowledge-based generalizations. arXiv:​2001.​05495
Zurück zum Zitat Baldini, I., Wei, D., Natesan Ramamurthy, K., et al. (2022). Your fairness may vary: Pretrained language model fairness in toxic text classification. In: Findings of the association for computational linguistics: ACL 2022. Association for Computational Linguistics, Dublin, Ireland, (pp. 2245–2262). https://doi.org/10.18653/v1/2022.findings-acl.176 Baldini, I., Wei, D., Natesan Ramamurthy, K., et al. (2022). Your fairness may vary: Pretrained language model fairness in toxic text classification. In: Findings of the association for computational linguistics: ACL 2022. Association for Computational Linguistics, Dublin, Ireland, (pp. 2245–2262). https://​doi.​org/​10.​18653/​v1/​2022.​findings-acl.​176
Zurück zum Zitat Borkan, D., Dixon, L., Sorensen, J., et al. (2019). Nuanced metrics for measuring unintended bias with real data for text classification. CoRR abs/1903.04561. arXiv:1903.04561 Borkan, D., Dixon, L., Sorensen, J., et al. (2019). Nuanced metrics for measuring unintended bias with real data for text classification. CoRR abs/1903.04561. arXiv:​1903.​04561
Zurück zum Zitat Chauhan, DS., S R, D., Ekbal, A., et al. (2020). Sentiment and emotion help sarcasm? a multi-task learning framework for multi-modal sarcasm, sentiment and emotion analysis. In: Proceedings of the 58th Annual meeting of the association for computational linguistics. Association for Computational Linguistics, Online, (pp. 4351–4360). https://doi.org/10.18653/v1/2020.acl-main.401 Chauhan, DS., S R, D., Ekbal, A., et al. (2020). Sentiment and emotion help sarcasm? a multi-task learning framework for multi-modal sarcasm, sentiment and emotion analysis. In: Proceedings of the 58th Annual meeting of the association for computational linguistics. Association for Computational Linguistics, Online, (pp. 4351–4360). https://​doi.​org/​10.​18653/​v1/​2020.​acl-main.​401
Zurück zum Zitat Cheng, L., Mosallanezhad, A., Silva, Y., et al. (2021). Mitigating bias in session-based cyberbullying detection: A non-compromising approach. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (vol. 1: Long Papers). Association for Computational Linguistics, Online, (pp. 2158–2168). https://doi.org/10.18653/v1/2021.acl-long.168 Cheng, L., Mosallanezhad, A., Silva, Y., et al. (2021). Mitigating bias in session-based cyberbullying detection: A non-compromising approach. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (vol. 1: Long Papers). Association for Computational Linguistics, Online, (pp. 2158–2168). https://​doi.​org/​10.​18653/​v1/​2021.​acl-long.​168
Zurück zum Zitat Conneau, A., Khandelwal, K., Goyal, N., et al. (2020). Unsupervised cross-lingual representation learning at scale. In: Proceedings of the 58th annual meeting of the association for computational linguistics. Association for Computational Linguistics, Online, (pp 8440–8451). https://doi.org/10.18653/v1/2020.acl-main.747 Conneau, A., Khandelwal, K., Goyal, N., et al. (2020). Unsupervised cross-lingual representation learning at scale. In: Proceedings of the 58th annual meeting of the association for computational linguistics. Association for Computational Linguistics, Online, (pp 8440–8451). https://​doi.​org/​10.​18653/​v1/​2020.​acl-main.​747
Zurück zum Zitat Dadvar, M., Trieschnigg, D., Ordelman, R., et al. (2013). Improving cyberbullying detection with user context. In: Proceedings of the 35th European conference on advances in information retrieval. Springer-Verlag, Berlin, Heidelberg, ECIR’13, (pp. 693–696). https://doi.org/10.1007/978-3-642-36973-5_62 Dadvar, M., Trieschnigg, D., Ordelman, R., et al. (2013). Improving cyberbullying detection with user context. In: Proceedings of the 35th European conference on advances in information retrieval. Springer-Verlag, Berlin, Heidelberg, ECIR’13, (pp. 693–696). https://​doi.​org/​10.​1007/​978-3-642-36973-5_​62
Zurück zum Zitat Davidson, T., Bhattacharya, D., Weber, I. (2019). Racial bias in hate speech and abusive language detection datasets. In: Proceedings of the third workshop on abusive language online. Association for Computational Linguistics, Florence, Italy, (pp. 25–35). https://doi.org/10.18653/v1/W19-3504 Davidson, T., Bhattacharya, D., Weber, I. (2019). Racial bias in hate speech and abusive language detection datasets. In: Proceedings of the third workshop on abusive language online. Association for Computational Linguistics, Florence, Italy, (pp. 25–35). https://​doi.​org/​10.​18653/​v1/​W19-3504
Zurück zum Zitat Dixon, L., Li, J., Sorensen, J., et al. (2018). Measuring and mitigating unintended bias in text classification. In: Proceedings of the 2018 AAAI/ACM conference on AI, ethics, and society. Association for Computing Machinery, New York, USA, AIES ’18, (pp. 67–73). https://doi.org/10.1145/3278721.3278729 Dixon, L., Li, J., Sorensen, J., et al. (2018). Measuring and mitigating unintended bias in text classification. In: Proceedings of the 2018 AAAI/ACM conference on AI, ethics, and society. Association for Computing Machinery, New York, USA, AIES ’18, (pp. 67–73). https://​doi.​org/​10.​1145/​3278721.​3278729
Zurück zum Zitat Godoy, D., & Tommasel, A. (2021). Is my model biased? exploring unintended bias in misogyny detection tasks. In: AIofAI 2021: 1st workshop on adverse impacts and collateral effects of artificial intelligence technologies, Montreal, Canada, (pp. 97–11) Godoy, D., & Tommasel, A. (2021). Is my model biased? exploring unintended bias in misogyny detection tasks. In: AIofAI 2021: 1st workshop on adverse impacts and collateral effects of artificial intelligence technologies, Montreal, Canada, (pp. 97–11)
Zurück zum Zitat Hada, R., Sudhir, S., Mishra, P., et al. (2021). Ruddit: Norms of offensiveness for English Reddit comments. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (vol. 1: Long Papers). Association for Computational Linguistics, Online, (pp. 2700–2717). https://doi.org/10.18653/v1/2021.acl-long.210 Hada, R., Sudhir, S., Mishra, P., et al. (2021). Ruddit: Norms of offensiveness for English Reddit comments. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (vol. 1: Long Papers). Association for Computational Linguistics, Online, (pp. 2700–2717). https://​doi.​org/​10.​18653/​v1/​2021.​acl-long.​210
Zurück zum Zitat Hardt, M., Price, E., & Srebro, N. (2016). Equality of opportunity in supervised learning. In: Proceedings of the 30th international conference on neural information processing systems. Curran Associates Inc., Red Hook, NY, USA, NIPS’16, (pp. 3323–3331) Hardt, M., Price, E., & Srebro, N. (2016). Equality of opportunity in supervised learning. In: Proceedings of the 30th international conference on neural information processing systems. Curran Associates Inc., Red Hook, NY, USA, NIPS’16, (pp. 3323–3331)
Zurück zum Zitat Hossain, E., Sharif, O., Hoque, M. M. (2022a). MUTE: A multimodal dataset for detecting hateful memes. In: Proceedings of the 2nd conference of the asia-pacific chapter of the association for computational linguistics and the 12th international joint conference on natural language processing: student research workshop. Association for Computational Linguistics, Online, (pp. 32–39). https://aclanthology.org/2022.aacl-srw.5 Hossain, E., Sharif, O., Hoque, M. M. (2022a). MUTE: A multimodal dataset for detecting hateful memes. In: Proceedings of the 2nd conference of the asia-pacific chapter of the association for computational linguistics and the 12th international joint conference on natural language processing: student research workshop. Association for Computational Linguistics, Online, (pp. 32–39). https://​aclanthology.​org/​2022.​aacl-srw.​5
Zurück zum Zitat Hu, A., & Flaxman, S. (2018). Multimodal sentiment analysis to explore the structure of emotions. In: Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery &; data mining. Association for Computing Machinery, New York, USA, KDD ’18, (pp. 350–358). https://doi.org/10.1145/3219819.3219853 Hu, A., & Flaxman, S. (2018). Multimodal sentiment analysis to explore the structure of emotions. In: Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery &; data mining. Association for Computing Machinery, New York, USA, KDD ’18, (pp. 350–358). https://​doi.​org/​10.​1145/​3219819.​3219853
Zurück zum Zitat Hukkelås, H., Mester, R., & Lindseth, F. (2019). Deepprivacy: A generative adversarial network for face anonymization. arXiv:1909.04538 Hukkelås, H., Mester, R., & Lindseth, F. (2019). Deepprivacy: A generative adversarial network for face anonymization. arXiv:​1909.​04538
Zurück zum Zitat Kiritchenko, S., & Mohammad, S. M. (2018). Examining gender and race bias in two hundred sentiment analysis systems. arXiv:1805.04508 Kiritchenko, S., & Mohammad, S. M. (2018). Examining gender and race bias in two hundred sentiment analysis systems. arXiv:​1805.​04508
Zurück zum Zitat Kumar, R., Ojha, A. K., Malmasi, S., et al. (2018). Benchmarking aggression identification in social media. In: Proceedings of the first workshop on trolling, aggression and cyberbullying (TRAC-2018). Association for Computational Linguistics, Santa Fe, New Mexico, USA, (pp. 1–11). https://aclanthology.org/W18-4401 Kumar, R., Ojha, A. K., Malmasi, S., et al. (2018). Benchmarking aggression identification in social media. In: Proceedings of the first workshop on trolling, aggression and cyberbullying (TRAC-2018). Association for Computational Linguistics, Santa Fe, New Mexico, USA, (pp. 1–11). https://​aclanthology.​org/​W18-4401
Zurück zum Zitat Kumar, S., Wintner, S., Smith, N. A., et al. (2019). Topics to avoid: Demoting latent confounds in text classification. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP). Association for Computational Linguistics, Hong Kong, China, (pp. 4153–4163). https://doi.org/10.18653/v1/D19-1425 Kumar, S., Wintner, S., Smith, N. A., et al. (2019). Topics to avoid: Demoting latent confounds in text classification. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP). Association for Computational Linguistics, Hong Kong, China, (pp. 4153–4163). https://​doi.​org/​10.​18653/​v1/​D19-1425
Zurück zum Zitat Liu, C., Geigle, G., Krebs, R., et al. (2022). FigMemes: A dataset for figurative language identification in politically-opinionated memes. In: Proceedings of the 2022 conference on empirical methods in natural language processing. Association for Computational Linguistics, Abu Dhabi, United Arab Emirates, (pp. 7069–7086). https://aclanthology.org/2022.emnlp-main.476 Liu, C., Geigle, G., Krebs, R., et al. (2022). FigMemes: A dataset for figurative language identification in politically-opinionated memes. In: Proceedings of the 2022 conference on empirical methods in natural language processing. Association for Computational Linguistics, Abu Dhabi, United Arab Emirates, (pp. 7069–7086). https://​aclanthology.​org/​2022.​emnlp-main.​476
Zurück zum Zitat Nobata, C., Tetreault, J., Thomas, A. et al. (2016). Abusive language detection in online user content. In: Proceedings of the 25th international conference on World Wide Web. International World Wide Web conferences steering committee, Republic and Canton of Geneva, CHE, WWW ’16, (pp. 145–153). https://doi.org/10.1145/2872427.2883062 Nobata, C., Tetreault, J., Thomas, A. et al. (2016). Abusive language detection in online user content. In: Proceedings of the 25th international conference on World Wide Web. International World Wide Web conferences steering committee, Republic and Canton of Geneva, CHE, WWW ’16, (pp. 145–153). https://​doi.​org/​10.​1145/​2872427.​2883062
Zurück zum Zitat Nozza, D. (2021). Exposing the limits of zero-shot cross-lingual hate speech detection. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (vol. 2: Short Papers). Association for Computational Linguistics, Online, (pp. 907–914). https://doi.org/10.18653/v1/2021.acl-short.114 Nozza, D. (2021). Exposing the limits of zero-shot cross-lingual hate speech detection. In: Proceedings of the 59th annual meeting of the association for computational linguistics and the 11th international joint conference on natural language processing (vol. 2: Short Papers). Association for Computational Linguistics, Online, (pp. 907–914). https://​doi.​org/​10.​18653/​v1/​2021.​acl-short.​114
Zurück zum Zitat Park, J. H., Shin, J., & Fung, P. (2018) Reducing gender bias in abusive language detection. In: Proceedings of the 2018 conference on empirical methods in natural language processing. Association for Computational Linguistics, Brussels, Belgium, (pp. 2799–2804). https://doi.org/10.18653/v1/D18-1302 Park, J. H., Shin, J., & Fung, P. (2018) Reducing gender bias in abusive language detection. In: Proceedings of the 2018 conference on empirical methods in natural language processing. Association for Computational Linguistics, Brussels, Belgium, (pp. 2799–2804). https://​doi.​org/​10.​18653/​v1/​D18-1302
Zurück zum Zitat Pramanick S, Sharma S, Dimitrov D, et al. (2021b) MOMENTA: A multimodal framework for detecting harmful memes and their targets. In: Findings of the association for computational linguistics: EMNLP 2021. Association for Computational Linguistics, Punta Cana, Dominican Republic, (pp. 4439–4455). https://doi.org/10.18653/v1/2021.findings-emnlp.379 Pramanick S, Sharma S, Dimitrov D, et al. (2021b) MOMENTA: A multimodal framework for detecting harmful memes and their targets. In: Findings of the association for computational linguistics: EMNLP 2021. Association for Computational Linguistics, Punta Cana, Dominican Republic, (pp. 4439–4455). https://​doi.​org/​10.​18653/​v1/​2021.​findings-emnlp.​379
Zurück zum Zitat Premjith, B., Chakravarthi, B. R., Subramanian, M., et al. (2022) Findings of the shared task on multimodal sentiment analysis and troll meme classification in Dravidian languages. In: Proceedings of the second workshop on speech and language technologies for Dravidian languages. Association for Computational Linguistics, Dublin, Ireland, (pp. 254–260). https://doi.org/10.18653/v1/2022.dravidianlangtech-1.39 Premjith, B., Chakravarthi, B. R., Subramanian, M., et al. (2022) Findings of the shared task on multimodal sentiment analysis and troll meme classification in Dravidian languages. In: Proceedings of the second workshop on speech and language technologies for Dravidian languages. Association for Computational Linguistics, Dublin, Ireland, (pp. 254–260). https://​doi.​org/​10.​18653/​v1/​2022.​dravidianlangtec​h-1.​39
Zurück zum Zitat Radford, A., Kim, J. W., Hallacy, C., et al. (2021). Learning transferable visual models from natural language supervision. In: M. Meila, T. Zhang (eds.), Proceedings of the 38th international conference on machine learning, proceedings of machine learning research, vol. 139. PMLR, (pp. 8748–8763). https://proceedings.mlr.press/v139/radford21a.html Radford, A., Kim, J. W., Hallacy, C., et al. (2021). Learning transferable visual models from natural language supervision. In: M. Meila, T. Zhang (eds.), Proceedings of the 38th international conference on machine learning, proceedings of machine learning research, vol. 139. PMLR, (pp. 8748–8763). https://​proceedings.​mlr.​press/​v139/​radford21a.​html
Zurück zum Zitat Ranasinghe, T., & Zampieri, M. (2020). Multilingual offensive language identification with cross-lingual embeddings. In: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP). Association for Computational Linguistics, Online, (pp. 5838–5844). https://doi.org/10.18653/v1/2020.emnlp-main.470 Ranasinghe, T., & Zampieri, M. (2020). Multilingual offensive language identification with cross-lingual embeddings. In: Proceedings of the 2020 conference on empirical methods in natural language processing (EMNLP). Association for Computational Linguistics, Online, (pp. 5838–5844). https://​doi.​org/​10.​18653/​v1/​2020.​emnlp-main.​470
Zurück zum Zitat Ribeiro, M.T., Singh, S., Guestrin, C. (2016). Why should I trust you?: Explaining the predictions of any classifier. arXiv:1602.04938 Ribeiro, M.T., Singh, S., Guestrin, C. (2016). Why should I trust you?: Explaining the predictions of any classifier. arXiv:​1602.​04938
Zurück zum Zitat Rijhwani, S., Sequiera, R., Choudhury, M., et al. (2017). Estimating code-switching on Twitter with a novel generalized word-level language detection technique. In: Proceedings of the 55th annual meeting of the association for computational linguistics (vol. 1: Long Papers). Association for Computational Linguistics, Vancouver, Canada, (pp. 1971–1982). https://doi.org/10.18653/v1/P17-1180 Rijhwani, S., Sequiera, R., Choudhury, M., et al. (2017). Estimating code-switching on Twitter with a novel generalized word-level language detection technique. In: Proceedings of the 55th annual meeting of the association for computational linguistics (vol. 1: Long Papers). Association for Computational Linguistics, Vancouver, Canada, (pp. 1971–1982). https://​doi.​org/​10.​18653/​v1/​P17-1180
Zurück zum Zitat Sharma, C., Bhageria, D., Scott, W., et al. (2020a). SemEval-2020 task 8: Memotion analysis- the visuo-lingual metaphor! In: Proceedings of the fourteenth workshop on semantic evaluation. International Committee for Computational Linguistics, Barcelona (online), (pp. 759–773). https://doi.org/10.18653/v1/2020.semeval-1.99 Sharma, C., Bhageria, D., Scott, W., et al. (2020a). SemEval-2020 task 8: Memotion analysis- the visuo-lingual metaphor! In: Proceedings of the fourteenth workshop on semantic evaluation. International Committee for Computational Linguistics, Barcelona (online), (pp. 759–773). https://​doi.​org/​10.​18653/​v1/​2020.​semeval-1.​99
Zurück zum Zitat Sharma, M., Kandasamy, I., Vasantha, W. (2020b). Memebusters at SemEval-2020 task 8: Feature fusion model for sentiment analysis on memes using transfer learning. In: A. Herbelot, X. Zhu, A. Palmer, et al. (eds.), Proceedings of the fourteenth workshop on semantic evaluation. International Committee for Computational Linguistics, Barcelona (online), (pp. 1163–1171). https://doi.org/10.18653/v1/2020.semeval-1.154 Sharma, M., Kandasamy, I., Vasantha, W. (2020b). Memebusters at SemEval-2020 task 8: Feature fusion model for sentiment analysis on memes using transfer learning. In: A. Herbelot, X. Zhu, A. Palmer, et al. (eds.), Proceedings of the fourteenth workshop on semantic evaluation. International Committee for Computational Linguistics, Barcelona (online), (pp. 1163–1171). https://​doi.​org/​10.​18653/​v1/​2020.​semeval-1.​154
Zurück zum Zitat Sharma, S., Siddiqui, M. K., Akhtar, M. S., et al. (2022b). Domain-aware self-supervised pre-training for label-efficient meme analysis. In: Proceedings of the 2nd conference of the Asia-Pacific chapter of the association for computational linguistics and the 12th international joint conference on natural language processing (vol. 1: Long Papers). Association for Computational Linguistics, Online only, (pp. 792–805). https://aclanthology.org/2022.aacl-main.60 Sharma, S., Siddiqui, M. K., Akhtar, M. S., et al. (2022b). Domain-aware self-supervised pre-training for label-efficient meme analysis. In: Proceedings of the 2nd conference of the Asia-Pacific chapter of the association for computational linguistics and the 12th international joint conference on natural language processing (vol. 1: Long Papers). Association for Computational Linguistics, Online only, (pp. 792–805). https://​aclanthology.​org/​2022.​aacl-main.​60
Zurück zum Zitat Simonyan, K., & Zisserman, A. (2015). Very deep convolutional networks for large-scale image recognition. Computational and Biological Learning Society, (pp. 1–14) Simonyan, K., & Zisserman, A. (2015). Very deep convolutional networks for large-scale image recognition. Computational and Biological Learning Society, (pp. 1–14)
Zurück zum Zitat Singh, P., & Lefever, E. (2021). LT3 at SemEval-2021 task 6: Using multi-modal compact bilinear pooling to combine visual and textual understanding in memes. In: Proceedings of the 15th international workshop on semantic evaluation (SemEval-2021). Association for Computational Linguistics, Online, (pp. 1051–1055). https://doi.org/10.18653/v1/2021.semeval-1.145 Singh, P., & Lefever, E. (2021). LT3 at SemEval-2021 task 6: Using multi-modal compact bilinear pooling to combine visual and textual understanding in memes. In: Proceedings of the 15th international workshop on semantic evaluation (SemEval-2021). Association for Computational Linguistics, Online, (pp. 1051–1055). https://​doi.​org/​10.​18653/​v1/​2021.​semeval-1.​145
Zurück zum Zitat Spinde, T., Rudnitckaia, L., Sinha, K., et al. (2021). MBIC - A media bias annotation dataset including annotator characteristics. arXiv:2105.11910 Spinde, T., Rudnitckaia, L., Sinha, K., et al. (2021). MBIC - A media bias annotation dataset including annotator characteristics. arXiv:​2105.​11910
Zurück zum Zitat Suryawanshi, S., Chakravarthi, B. R., Arcan, M., et al. (2020). Multimodal meme dataset (MultiOFF) for identifying offensive content in image and text. In: Proceedings of the second workshop on trolling, aggression and cyberbullying. European Language Resources Association (ELRA), Marseille, France, (pp. 32–41). https://aclanthology.org/2020.trac-1.6 Suryawanshi, S., Chakravarthi, B. R., Arcan, M., et al. (2020). Multimodal meme dataset (MultiOFF) for identifying offensive content in image and text. In: Proceedings of the second workshop on trolling, aggression and cyberbullying. European Language Resources Association (ELRA), Marseille, France, (pp. 32–41). https://​aclanthology.​org/​2020.​trac-1.​6
Zurück zum Zitat Tan, H., & Bansal, M. (2019). LXMERT: Learning cross-modality encoder representations from transformers. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP). Association for Computational Linguistics, Hong Kong, China, (pp. 5100–5111). https://doi.org/10.18653/v1/D19-1514 Tan, H., & Bansal, M. (2019). LXMERT: Learning cross-modality encoder representations from transformers. In: Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP). Association for Computational Linguistics, Hong Kong, China, (pp. 5100–5111). https://​doi.​org/​10.​18653/​v1/​D19-1514
Zurück zum Zitat Wiegand, M., Ruppenhofer, J., & Kleinbauer, T. (2019a). Detection of abusive language: The problem of biased datasets. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: Human Language Technologies, vol. 1 (Long and Short Papers). Association for Computational Linguistics, Minneapolis, Minnesota, (pp. 602–608). https://doi.org/10.18653/v1/N19-1060 Wiegand, M., Ruppenhofer, J., & Kleinbauer, T. (2019a). Detection of abusive language: The problem of biased datasets. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: Human Language Technologies, vol. 1 (Long and Short Papers). Association for Computational Linguistics, Minneapolis, Minnesota, (pp. 602–608). https://​doi.​org/​10.​18653/​v1/​N19-1060
Zurück zum Zitat Wiegand, M., Siegel, M., & Ruppenhofer, J. (2019b). Overview of the germeval 2018 shared task on the identification of offensive language. Proceedings of GermEval 2018, 14th conference on natural language processing (KONVENS 2018), Vienna, Austria, Austrian Academy of Sciences, Vienna, Austria, (pp. 1 – 10). Accessed 21 Sept 2018. https://nbn-resolving.org/urn:nbn:de:bsz:mh39-84935 Wiegand, M., Siegel, M., & Ruppenhofer, J. (2019b). Overview of the germeval 2018 shared task on the identification of offensive language. Proceedings of GermEval 2018, 14th conference on natural language processing (KONVENS 2018), Vienna, Austria, Austrian Academy of Sciences, Vienna, Austria, (pp. 1 – 10). Accessed 21 Sept 2018. https://​nbn-resolving.​org/​urn:​nbn:​de:​bsz:​mh39-84935
Zurück zum Zitat Xu, J. M., Jun, K. S., Zhu, X., et al. (2012). Learning from bullying traces in social media. In: Proceedings of the 2012 conference of the North American chapter of the association for computational linguistics: human language technologies. Association for Computational Linguistics, Montréal, Canada, (pp. 656–666). https://aclanthology.org/N12-1084 Xu, J. M., Jun, K. S., Zhu, X., et al. (2012). Learning from bullying traces in social media. In: Proceedings of the 2012 conference of the North American chapter of the association for computational linguistics: human language technologies. Association for Computational Linguistics, Montréal, Canada, (pp. 656–666). https://​aclanthology.​org/​N12-1084
Zurück zum Zitat Yu, Z., Yu, J., Fan, J., et al. (2017). Multi-modal factorized bilinear pooling with co-attention learning for visual question answering. In: 2017 IEEE International conference on computer vision (ICCV). IEEE Computer Society, Los Alamitos, CA, USA, (pp. 1839–1848). https://doi.org/10.1109/ICCV.2017.202 Yu, Z., Yu, J., Fan, J., et al. (2017). Multi-modal factorized bilinear pooling with co-attention learning for visual question answering. In: 2017 IEEE International conference on computer vision (ICCV). IEEE Computer Society, Los Alamitos, CA, USA, (pp. 1839–1848). https://​doi.​org/​10.​1109/​ICCV.​2017.​202
Zurück zum Zitat Zampieri, M., Malmasi, S., Nakov, P., et al. (2019a) Predicting the type and target of offensive posts in social media. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: Human Language Technologies, vol. 1 (Long and Short Papers). Association for Computational Linguistics, Minneapolis, Minnesota, (pp. 1415–1420). https://doi.org/10.18653/v1/N19-1144 Zampieri, M., Malmasi, S., Nakov, P., et al. (2019a) Predicting the type and target of offensive posts in social media. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: Human Language Technologies, vol. 1 (Long and Short Papers). Association for Computational Linguistics, Minneapolis, Minnesota, (pp. 1415–1420). https://​doi.​org/​10.​18653/​v1/​N19-1144
Zurück zum Zitat Zampieri, M., Malmasi, S., Nakov, P., et al. (2019b) SemEval-2019 task 6: Identifying and categorizing offensive language in social media (OffensEval). In: Proceedings of the 13th international workshop on semantic evaluation. Association for Computational Linguistics, Minneapolis, Minnesota, USA, (pp. 75–86). https://doi.org/10.18653/v1/S19-2010 Zampieri, M., Malmasi, S., Nakov, P., et al. (2019b) SemEval-2019 task 6: Identifying and categorizing offensive language in social media (OffensEval). In: Proceedings of the 13th international workshop on semantic evaluation. Association for Computational Linguistics, Minneapolis, Minnesota, USA, (pp. 75–86). https://​doi.​org/​10.​18653/​v1/​S19-2010
Zurück zum Zitat Zhou, K., Lai, E., & Jiang, J. (2022). VLStereoSet: A study of stereotypical bias in pre-trained vision-language models. In: Proceedings of the 2nd conference of the Asia-Pacific chapter of the association for computational linguistics and the 12th international joint conference on natural language processing (vol. 1: Long Papers). Association for Computational Linguistics, Online only, (pp. 527–538). https://aclanthology.org/2022.aacl-main.40 Zhou, K., Lai, E., & Jiang, J. (2022). VLStereoSet: A study of stereotypical bias in pre-trained vision-language models. In: Proceedings of the 2nd conference of the Asia-Pacific chapter of the association for computational linguistics and the 12th international joint conference on natural language processing (vol. 1: Long Papers). Association for Computational Linguistics, Online only, (pp. 527–538). https://​aclanthology.​org/​2022.​aacl-main.​40
Zurück zum Zitat Zhou, L., Palangi, H., Zhang, L., et al. (2020). Unified vision-language pre-training for image captioning and vqa. Proceedings of the AAAI Conference on Artificial Intelligence,34(07), 13041–13049. Zhou, L., Palangi, H., Zhang, L., et al. (2020). Unified vision-language pre-training for image captioning and vqa. Proceedings of the AAAI Conference on Artificial Intelligence,34(07), 13041–13049.
Metadaten
Titel
Enhancing the fairness of offensive memes detection models by mitigating unintended political bias
verfasst von
Gitanjali Kumari
Anubhav Sinha
Asif Ekbal
Arindam Chatterjee
Vinutha B N
Publikationsdatum
06.01.2024
Verlag
Springer US
Erschienen in
Journal of Intelligent Information Systems
Print ISSN: 0925-9902
Elektronische ISSN: 1573-7675
DOI
https://doi.org/10.1007/s10844-023-00834-9