Skip to main content
Erschienen in: World Wide Web 6/2023

18.10.2023

CoSP: co-selection pick for a global explainability of black box machine learning models

verfasst von: Dou El Kefel Mansouri, Seif-Eddine Benkabou, Khaoula Meddahi, Allel Hadjali, Amin Mesmoudi, Khalid Benabdeslem, Souleyman Chaib

Erschienen in: World Wide Web | Ausgabe 6/2023

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

Recently, few methods for understanding machine learning model’s outputs have been developed. SHAP and LIME are two well-known examples of these methods. They provide individual explanations based on feature importance for each instance. While remarkable scores have been achieved for individual explanations, understanding the model’s decisions globally remains a complex task. Methods like LIME were extended to face this complexity by using individual explanations. In this approach, the problem was expressed as a submodular optimization problem. This algorithm is a bottom-up method aiming at providing a global explanation. It consists of picking a group of individual explanations which illustrate the global behavior of the model and avoid redundancy. In this paper, we propose CoSP (Co-Selection Pick) framework that allows a global explainability of any black-box model by selecting individual explanations based on a similarity preserving approach. Unlike submodular optimization, in our method the problem is considered as a co-selection task. This approach achieves a co-selection of instances and features over the explanations provided by any explainer. The proposed framework is more generic given that it is possible to make the co-selection either in supervised or unsupervised scenarios and also over explanations provided by any local explainer. Preliminary experimental results are made to validate our proposal.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Literatur
1.
Zurück zum Zitat Mohaghegh, F., Murthy, J.: Machine learning and computer vision techniques to predict thermal properties of particulate composites. CoRR. abs/2010.01968 (2020). arXiv:2010.01968 Mohaghegh, F., Murthy, J.: Machine learning and computer vision techniques to predict thermal properties of particulate composites. CoRR. abs/2010.01968 (2020). arXiv:​2010.​01968
6.
Zurück zum Zitat Gunning, D., Aha, D.: Darpas explainable artificial intelligence (xai) program. AI. Mag. 40(2), 44–58 (2019) Gunning, D., Aha, D.: Darpas explainable artificial intelligence (xai) program. AI. Mag. 40(2), 44–58 (2019)
8.
Zurück zum Zitat Strumbelj, E., Kononenko, I.: Explaining prediction models and individual predictions with feature contributions. Knowl. Inf. Syst. 41, 647–665 (2013)CrossRef Strumbelj, E., Kononenko, I.: Explaining prediction models and individual predictions with feature contributions. Knowl. Inf. Syst. 41, 647–665 (2013)CrossRef
9.
Zurück zum Zitat Lundberg, S., Lee, S.: A unified approach to interpreting model predictions. In: Advances in Neural Information Processing Systems, pp. 4765–4774 (2017) Lundberg, S., Lee, S.: A unified approach to interpreting model predictions. In: Advances in Neural Information Processing Systems, pp. 4765–4774 (2017)
10.
Zurück zum Zitat Lundberg, S., Erion, G., Chen, H., DeGrave, A., Prutkin, J., Nair, B., Katz, R., Himmelfarb, J., Bansal, N., Lee, S.: Explainable ai for trees: From local explanations to global understanding. ArXiv. abs/1905.04610, (2019) Lundberg, S., Erion, G., Chen, H., DeGrave, A., Prutkin, J., Nair, B., Katz, R., Himmelfarb, J., Bansal, N., Lee, S.: Explainable ai for trees: From local explanations to global understanding. ArXiv. abs/1905.04610, (2019)
11.
Zurück zum Zitat Ribeiro, M., Singh, S., Guestrin, C.: “why should I trust you?”: Explaining the predictions of any classifier. In: al., B.K. (ed.) Proc. of the 22nd ACM SIGKDD Inter. Conf. on Knowledge Discovery and Data Mining, San Francisco, CA, USA, August 13-17, 2016, pp. 1135–1144. ACM, ??? (2016). https://doi.org/10.1145/2939672.2939778 Ribeiro, M., Singh, S., Guestrin, C.: “why should I trust you?”: Explaining the predictions of any classifier. In: al., B.K. (ed.) Proc. of the 22nd ACM SIGKDD Inter. Conf. on Knowledge Discovery and Data Mining, San Francisco, CA, USA, August 13-17, 2016, pp. 1135–1144. ACM, ??? (2016). https://​doi.​org/​10.​1145/​2939672.​2939778
12.
Zurück zum Zitat Li, J., Cheng, K., Wang, S., Morstatter, F., Trevino, R.P., Tang, J., Liu, H.: Feature selection: A data perspective. ACM computing surveys (CSUR). 50(6), 1–45 (2017)CrossRef Li, J., Cheng, K., Wang, S., Morstatter, F., Trevino, R.P., Tang, J., Liu, H.: Feature selection: A data perspective. ACM computing surveys (CSUR). 50(6), 1–45 (2017)CrossRef
13.
Zurück zum Zitat Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3(Mar), 1157–1182 (2003) Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3(Mar), 1157–1182 (2003)
14.
Zurück zum Zitat Olvera-López, J.A., Carrasco-Ochoa, J.A., Martínez-Trinidad, J.F., Kittler, J.: A review of instance selection methods. Artif. Intell. Rev. 34, 133–143 (2010)CrossRef Olvera-López, J.A., Carrasco-Ochoa, J.A., Martínez-Trinidad, J.F., Kittler, J.: A review of instance selection methods. Artif. Intell. Rev. 34, 133–143 (2010)CrossRef
16.
Zurück zum Zitat Li, Y.-F., Zhou, Z.-H.: Improving semi-supervised support vector machines through unlabeled instances selection. In: Proceedings of the AAAI Conference on Artificial Intelligence, 25, pp. 386–391 (2011) Li, Y.-F., Zhou, Z.-H.: Improving semi-supervised support vector machines through unlabeled instances selection. In: Proceedings of the AAAI Conference on Artificial Intelligence, 25, pp. 386–391 (2011)
17.
Zurück zum Zitat Kuncheva, L.I., Jain, L.C.: Nearest neighbor classifier: Simultaneous editing and feature selection. Pattern Recogn. Lett. 20(11–13), 1149–1156 (1999)CrossRef Kuncheva, L.I., Jain, L.C.: Nearest neighbor classifier: Simultaneous editing and feature selection. Pattern Recogn. Lett. 20(11–13), 1149–1156 (1999)CrossRef
18.
Zurück zum Zitat Derrac, J., García, S., Herrera, F.: Ifs-coco: Instance and feature selection based on cooperative coevolution with nearest neighbor rule. Pattern Recogn. 43(6), 2082–2105 (2010)CrossRef Derrac, J., García, S., Herrera, F.: Ifs-coco: Instance and feature selection based on cooperative coevolution with nearest neighbor rule. Pattern Recogn. 43(6), 2082–2105 (2010)CrossRef
19.
Zurück zum Zitat GarcíA-Pedrajas, N., De Haro-GarcíA, A., PéRez-RodríGuez, J.: A scalable approach to simultaneous evolutionary instance and feature selection. Inf. Sci. 228, 150–174 (2013)MathSciNetCrossRef GarcíA-Pedrajas, N., De Haro-GarcíA, A., PéRez-RodríGuez, J.: A scalable approach to simultaneous evolutionary instance and feature selection. Inf. Sci. 228, 150–174 (2013)MathSciNetCrossRef
20.
Zurück zum Zitat Zhao, Z., Wang, L., Liu, H., Ye, J.: On similarity preserving feature selection. IEEE Trans. Knowl. Data Eng. 25(3), 619–632 (2011)CrossRef Zhao, Z., Wang, L., Liu, H., Ye, J.: On similarity preserving feature selection. IEEE Trans. Knowl. Data Eng. 25(3), 619–632 (2011)CrossRef
21.
Zurück zum Zitat Ma, Y., Xu, X., Shen, F., Shen, H.T.: Similarity preserving feature generating networks for zero-shot learning. Neurocomputing 406, 333–342 (2020)CrossRef Ma, Y., Xu, X., Shen, F., Shen, H.T.: Similarity preserving feature generating networks for zero-shot learning. Neurocomputing 406, 333–342 (2020)CrossRef
22.
Zurück zum Zitat Shang, R., Chang, J., Jiao, L., Xue, Y.: Unsupervised feature selection based on self-representation sparse regression and local similarity preserving. Int. J. Mach. Learn. & Cybernet. 10, 757–770 (2019)CrossRef Shang, R., Chang, J., Jiao, L., Xue, Y.: Unsupervised feature selection based on self-representation sparse regression and local similarity preserving. Int. J. Mach. Learn. & Cybernet. 10, 757–770 (2019)CrossRef
24.
Zurück zum Zitat Meddahi, K., Benkabou, S.-E., Hadjali, A., Mesmoudi, A., El Kefel Mansouri, D., Benabdeslem, K., Chaib, S.: Towards a co-selection approach for a global explainability of black box machine learning models. In: International Conference on Web Information Systems Engineering, pp. 97–109 (2022). Springer Meddahi, K., Benkabou, S.-E., Hadjali, A., Mesmoudi, A., El Kefel Mansouri, D., Benabdeslem, K., Chaib, S.: Towards a co-selection approach for a global explainability of black box machine learning models. In: International Conference on Web Information Systems Engineering, pp. 97–109 (2022). Springer
25.
Zurück zum Zitat Martens, D., Provost, F.: Explaining data-driven document classifications. MIS quarterly. 38(1), 73–100 (2014) Martens, D., Provost, F.: Explaining data-driven document classifications. MIS quarterly. 38(1), 73–100 (2014)
26.
Zurück zum Zitat Baehrens, D., Schroeter, T., Harmeling, S., Kawanabe, M., Hansen, K., Müller, K.-R.: How to explain individual classification decisions. J. Mach. Learn. Res. 11, 1803–1831 (2010)MathSciNet Baehrens, D., Schroeter, T., Harmeling, S., Kawanabe, M., Hansen, K., Müller, K.-R.: How to explain individual classification decisions. J. Mach. Learn. Res. 11, 1803–1831 (2010)MathSciNet
28.
Zurück zum Zitat Burkart, N., Huber, M.F.: A survey on the explainability of supervised machine learning. J. Artif. Intell. Res. 70, 245–317 (2021)MathSciNetCrossRef Burkart, N., Huber, M.F.: A survey on the explainability of supervised machine learning. J. Artif. Intell. Res. 70, 245–317 (2021)MathSciNetCrossRef
29.
Zurück zum Zitat Lundberg, S.M., Erion, G., Chen, H., DeGrave, A., Prutkin, J.M., Nair, B., Katz, R., Himmelfarb, J., Bansal, N., Lee, S.-I.: From local explanations to global understanding with explainable ai for trees. Nat. Mach. Intell. 2(1), 56–67 (2020)CrossRef Lundberg, S.M., Erion, G., Chen, H., DeGrave, A., Prutkin, J.M., Nair, B., Katz, R., Himmelfarb, J., Bansal, N., Lee, S.-I.: From local explanations to global understanding with explainable ai for trees. Nat. Mach. Intell. 2(1), 56–67 (2020)CrossRef
30.
Zurück zum Zitat Vlahek, D., Mongus, D.: An efficient iterative approach to explainable feature learning. IEEE Trans. Neural Netw. & Learn, Syst (2021) Vlahek, D., Mongus, D.: An efficient iterative approach to explainable feature learning. IEEE Trans. Neural Netw. & Learn, Syst (2021)
31.
Zurück zum Zitat Dinh, V.C., Ho, L.S.: Consistent feature selection for analytic deep neural networks. Adv Neural Inf. Proc. Syst. 33, 2420–2431 (2020) Dinh, V.C., Ho, L.S.: Consistent feature selection for analytic deep neural networks. Adv Neural Inf. Proc. Syst. 33, 2420–2431 (2020)
32.
Zurück zum Zitat Cancela, B., Bolón-Canedo, V., Alonso-Betanzos, A.: E2e-fs: An end-to-end feature selection method for neural networks. IEEE Trans. Pattern Anal. & Mach, Intell (2022) Cancela, B., Bolón-Canedo, V., Alonso-Betanzos, A.: E2e-fs: An end-to-end feature selection method for neural networks. IEEE Trans. Pattern Anal. & Mach, Intell (2022)
33.
Zurück zum Zitat Wang, X., Wu, Y., Zhang, A., Feng, F., He, X., Chua, T.-S.: Reinforced causal explainer for graph neural networks. IEEE Trans. Pattern. Anal. & Mach, Intell (2022) Wang, X., Wu, Y., Zhang, A., Feng, F., He, X., Chua, T.-S.: Reinforced causal explainer for graph neural networks. IEEE Trans. Pattern. Anal. & Mach, Intell (2022)
34.
Zurück zum Zitat Böhle, M., Fritz, M., Schiele, B.: Optimising for interpretability: Convolutional dynamic alignment networks. IEEE Transactions on Pattern Analysis and Machine Intelligence. (2022) Böhle, M., Fritz, M., Schiele, B.: Optimising for interpretability: Convolutional dynamic alignment networks. IEEE Transactions on Pattern Analysis and Machine Intelligence. (2022)
35.
Zurück zum Zitat Shrikumar, A., Greenside, P., Shcherbina, A., Kundaje, A.: Not just a black box: Learning important features through propagating activation differences. arXiv:1605.01713 (2016) Shrikumar, A., Greenside, P., Shcherbina, A., Kundaje, A.: Not just a black box: Learning important features through propagating activation differences. arXiv:​1605.​01713 (2016)
36.
Zurück zum Zitat Krishnan, S., Wu, E.: Palm: Machine learning explanations for iterative debugging. In: Proceedings of the 2Nd Workshop on Human-in-the-loop Data Analytics, pp. 1–6 (2017) Krishnan, S., Wu, E.: Palm: Machine learning explanations for iterative debugging. In: Proceedings of the 2Nd Workshop on Human-in-the-loop Data Analytics, pp. 1–6 (2017)
37.
Zurück zum Zitat Ribeiro, M.T., Singh, S., Guestrin, C.: Anchors: High-precision model-agnostic explanations. In: Proc. AAAI Conf. Artif. Intell. vol. 32 (2018) Ribeiro, M.T., Singh, S., Guestrin, C.: Anchors: High-precision model-agnostic explanations. In: Proc. AAAI Conf. Artif. Intell. vol. 32 (2018)
38.
Zurück zum Zitat Zhou, B., Bau, D., Oliva, A., Torralba, A.: Interpreting deep visual representations via network dissection. IEEE Trans. Pattern. Anal. & Mach. Intell. 41(9), 2131–2145 (2018) Zhou, B., Bau, D., Oliva, A., Torralba, A.: Interpreting deep visual representations via network dissection. IEEE Trans. Pattern. Anal. & Mach. Intell. 41(9), 2131–2145 (2018)
39.
Zurück zum Zitat Schwab, P., Karlen, W.: Cxplain: Causal explanations for model interpretation under uncertainty. Adv. Neural Inf. Proc. Syst. 32 (2019) Schwab, P., Karlen, W.: Cxplain: Causal explanations for model interpretation under uncertainty. Adv. Neural Inf. Proc. Syst. 32 (2019)
41.
Zurück zum Zitat Albini, E., Rago, A., Baroni, P., Toni, F.: Relation-based counterfactual explanations for bayesian network classifiers. In: IJCAI, pp. 451–457 (2020) Albini, E., Rago, A., Baroni, P., Toni, F.: Relation-based counterfactual explanations for bayesian network classifiers. In: IJCAI, pp. 451–457 (2020)
42.
Zurück zum Zitat Agarwal, R., Melnick, L., Frosst, N., Zhang, X., Lengerich, B., Caruana, R., Hinton, G.E.: Neural additive models: Interpretable machine learning with neural nets. Adv. Neural Inf. Proc. Syst. 34, 4699–4711 (2021) Agarwal, R., Melnick, L., Frosst, N., Zhang, X., Lengerich, B., Caruana, R., Hinton, G.E.: Neural additive models: Interpretable machine learning with neural nets. Adv. Neural Inf. Proc. Syst. 34, 4699–4711 (2021)
43.
Zurück zum Zitat Liu, Y., Khandagale, S., White, C., Neiswanger, W.: Synthetic benchmarks for scientific research in explainable machine learning. In: Adv. Neural Inf. Proc. Syst. Datasets Track. (2021) Liu, Y., Khandagale, S., White, C., Neiswanger, W.: Synthetic benchmarks for scientific research in explainable machine learning. In: Adv. Neural Inf. Proc. Syst. Datasets Track. (2021)
44.
Zurück zum Zitat Graziani, M., Lompech, T., Müller, H., Andrearczyk, V.: Evaluation and comparison of cnn visual explanations for histopathology. In: Proceedings of the AAAI Conference on Artificial Intelligence Workshops (XAI-AAAI-21), Virtual Event, pp. 8–9 (2021) Graziani, M., Lompech, T., Müller, H., Andrearczyk, V.: Evaluation and comparison of cnn visual explanations for histopathology. In: Proceedings of the AAAI Conference on Artificial Intelligence Workshops (XAI-AAAI-21), Virtual Event, pp. 8–9 (2021)
45.
Zurück zum Zitat Agarwal, C., Krishna, S., Saxena, E., Pawelczyk, M., Johnson, N., Puri, I., Zitnik, M., Lakkaraju, H.: Openxai: Towards a transparent evaluation of model explanations. Adv. Neural Inf. Proc. Syst. 35, 15784–15799 (2022) Agarwal, C., Krishna, S., Saxena, E., Pawelczyk, M., Johnson, N., Puri, I., Zitnik, M., Lakkaraju, H.: Openxai: Towards a transparent evaluation of model explanations. Adv. Neural Inf. Proc. Syst. 35, 15784–15799 (2022)
46.
Zurück zum Zitat Li, X., Xiong, H., Li, X., Wu, X., Chen, Z., Dou, D.: Interpretdl: explaining deep models in paddlepaddle. J. Mach. Learn. Res. 23(1), 8969–8974 (2022)MathSciNet Li, X., Xiong, H., Li, X., Wu, X., Chen, Z., Dou, D.: Interpretdl: explaining deep models in paddlepaddle. J. Mach. Learn. Res. 23(1), 8969–8974 (2022)MathSciNet
47.
Zurück zum Zitat Motallebi, M., Anik, M.T.A., Zaïane, O.R.: Explaining decisions of black-box models using barbe. In: Int. Conf. Database & Expert Syst. Appl. pp. 82–97 (2023). Springer Motallebi, M., Anik, M.T.A., Zaïane, O.R.: Explaining decisions of black-box models using barbe. In: Int. Conf. Database & Expert Syst. Appl. pp. 82–97 (2023). Springer
48.
Zurück zum Zitat Agarwal, C., Queen, O., Lakkaraju, H., Zitnik, M.: Evaluating explainability for graph neural networks. Scientific Data. 10(1), 144 (2023)CrossRef Agarwal, C., Queen, O., Lakkaraju, H., Zitnik, M.: Evaluating explainability for graph neural networks. Scientific Data. 10(1), 144 (2023)CrossRef
50.
51.
Zurück zum Zitat Tong, H., Lin, C.: Non-negative residual matrix factorization with application to graph anomaly detection. In: Proceedings of the Eleventh SIAM International Conference on Data Mining, SDM 2011, April 28-30, 2011, Mesa, Arizona, USA, pp. 143–153. SIAM / Omnipress, ??? (2011) Tong, H., Lin, C.: Non-negative residual matrix factorization with application to graph anomaly detection. In: Proceedings of the Eleventh SIAM International Conference on Data Mining, SDM 2011, April 28-30, 2011, Mesa, Arizona, USA, pp. 143–153. SIAM / Omnipress, ??? (2011)
52.
Zurück zum Zitat Tang, J., Liu, H.: Coselect: Feature selection with instance selection for social media data. In: Proceedings of the 13th SIAM International Conference on Data Mining, May 2-4, 2013. Austin, Texas, USA, pp. 695–703. SIAM, ??? (2013) Tang, J., Liu, H.: Coselect: Feature selection with instance selection for social media data. In: Proceedings of the 13th SIAM International Conference on Data Mining, May 2-4, 2013. Austin, Texas, USA, pp. 695–703. SIAM, ??? (2013)
53.
Zurück zum Zitat Minaee, S., Kalchbrenner, N., Cambria, E., Nikzad, N., Chenaghlu, M., Gao, J.: Deep learning-based text classification. ACM Computing Surveys (CSUR). 54, 1–40 (2021)CrossRef Minaee, S., Kalchbrenner, N., Cambria, E., Nikzad, N., Chenaghlu, M., Gao, J.: Deep learning-based text classification. ACM Computing Surveys (CSUR). 54, 1–40 (2021)CrossRef
54.
Zurück zum Zitat Linden, I.-V.-D., Haned, H., Kanoulas, E.: Global aggregations of local explanations for black box models. CoRR. abs/1907.03039, (2019). arXiv:1907.03039 Linden, I.-V.-D., Haned, H., Kanoulas, E.: Global aggregations of local explanations for black box models. CoRR. abs/1907.03039, (2019). arXiv:​1907.​03039
Metadaten
Titel
CoSP: co-selection pick for a global explainability of black box machine learning models
verfasst von
Dou El Kefel Mansouri
Seif-Eddine Benkabou
Khaoula Meddahi
Allel Hadjali
Amin Mesmoudi
Khalid Benabdeslem
Souleyman Chaib
Publikationsdatum
18.10.2023
Verlag
Springer US
Erschienen in
World Wide Web / Ausgabe 6/2023
Print ISSN: 1386-145X
Elektronische ISSN: 1573-1413
DOI
https://doi.org/10.1007/s11280-023-01213-8

Weitere Artikel der Ausgabe 6/2023

World Wide Web 6/2023 Zur Ausgabe

Premium Partner