skip to main content
survey

Explainable AI (XAI): Core Ideas, Techniques, and Solutions

Published:16 January 2023Publication History
Skip Abstract Section

Abstract

As our dependence on intelligent machines continues to grow, so does the demand for more transparent and interpretable models. In addition, the ability to explain the model generally is now the gold standard for building trust and deployment of artificial intelligence systems in critical domains. Explainable artificial intelligence (XAI) aims to provide a suite of machine learning techniques that enable human users to understand, appropriately trust, and produce more explainable models. Selecting an appropriate approach for building an XAI-enabled application requires a clear understanding of the core ideas within XAI and the associated programming frameworks. We survey state-of-the-art programming techniques for XAI and present the different phases of XAI in a typical machine learning development process. We classify the various XAI approaches and, using this taxonomy, discuss the key differences among the existing XAI techniques. Furthermore, concrete examples are used to describe these techniques that are mapped to programming frameworks and software toolkits. It is the intention that this survey will help stakeholders in selecting the appropriate approaches, programming frameworks, and software toolkits by comparing them through the lens of the presented taxonomy.

REFERENCES

  1. [1] Angelov Plamen and Soares Eduardo. 2020. Towards explainable deep neural networks (xDNN). Neural Networks 130 (2020), 185194. Google ScholarGoogle ScholarCross RefCross Ref
  2. [2] Apley Daniel W. and Zhu Jingyu. 2016. Visualizing the effects of predictor variables in black box supervised learning models. arXiv preprint arXiv:1612.08468.Google ScholarGoogle Scholar
  3. [3] Bader Sebastian and Hitzler Pascal. 2005. Dimensions of neural-symbolic integration— structured survey. In We Will Show Them! Essays in Honour of Dov Gabbay, Volume One, Artëmov Sergei N., Barringer Howard, Garcez Artur S. d’Avila, Lamb Luís C., and Woods John (Eds.). College Publications, 167194.Google ScholarGoogle Scholar
  4. [4] Arrieta Alejandro Barredo, Díaz-Rodríguez Natalia, Ser Javier Del, Bennetot Adrien, Tabik Siham, Barbado Alberto, Garcia Salvador, et al. 2020. Explainable artificial intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion 58 (2020), 82115. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. [5] Beckett David, Berners-Lee Tim, Prud’hommeaux Eric, and Carothers Gavin. 2014. RDF 1.1 Turtle. World Wide Web Consortium 2014 (2014), 1831.Google ScholarGoogle Scholar
  6. [6] Besold Tarek R., Garcez Artur S. d’Avila, Bader Sebastian, Bowman Howard, Domingos Pedro M., Hitzler Pascal, Kühnberger Kai-Uwe, et al. 2017. Neural-symbolic learning and reasoning: A survey and interpretation. CoRR abs/1711.03902.Google ScholarGoogle Scholar
  7. [7] Bhatt Umang, Xiang Alice, Sharma Shubham, Weller Adrian, Taly Ankur, Jia Yunhan, Ghosh Joydeep, Puri Ruchir, Moura José M. F., and Eckersley Peter. 2020. Explainable machine learning in deployment. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency.ACM, New York, NY, 648657. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. [8] Bühmann Lorenz, Lehmann Jens, and Westphal Patrick. 2016. DL-Learner—A framework for inductive learning on the semantic web. Journal of Web Semantics 39 (2016), 1524.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. [9] Carvalho Diogo V., Pereira Eduardo M., and Cardoso Jaime S.. 2019. Machine learning interpretability: A survey on methods and metrics. Electronics 8, 8 (2019), 832.Google ScholarGoogle ScholarCross RefCross Ref
  10. [10] Corporation. NVIDIA2015. NVIDIA Collective Communications Library (NCCL). Retrieved July 15, 2021 from https://developer.nvidia.com/nccl.Google ScholarGoogle Scholar
  11. [11] Dean Jeffrey, Corrado Greg S., Monga Rajat, Chen Kai, Devin Matthieu, Le Quoc V., Mao Mark Z., et al. 2012. Large scale distributed deep networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems (NIPS’12). 1223–1231.Google ScholarGoogle Scholar
  12. [12] Dhurandhar Amit, Chen Pin-Yu, Luss Ronny, Tu Chun-Chen, Ting Paishun, Shanmugam Karthikeyan, and Das Payel. 2018. Explanations based on the missing: Towards contrastive explanations with pertinent negatives. In Advances in Neural Information Processing Systems. 592603.Google ScholarGoogle Scholar
  13. [13] Doran Derek, Schulz Sarah, and Besold Tarek R.. 2017. What does explainable AI really mean? A new conceptualization of perspectives. In Proceedings of the First International Workshop on Comprehensibility and Explanation in AI and ML 2017 co-located with 16th International Conference of the Italian Association for Artificial Intelligence (AI*IA 2017), Bari, Italy, November 16th and 17th, 2017(CEUR Workshop Proceedings, Vol. 2071), Besold Tarek R. and Kutz Oliver (Eds.). CEUR-WS.org. http://ceur-ws.org/Vol-2071/CExAIIA_2017_paper_2.pdf.Google ScholarGoogle Scholar
  14. [14] Doshi-Velez Finale and Kim Been. 2017. Towards a rigorous science of interpretable machine learning. arxiv:1702.08608 [stat.ML].Google ScholarGoogle Scholar
  15. [15] Došilović Filip Karlo, Brčić Mario, and Hlupić Nikica. 2018. Explainable artificial intelligence: A survey. In Proceedings of the 2018 41st International Convention on Information and Communication Technology, Electronics, and Microelectronics (MIPRO’18). 02100215. Google ScholarGoogle ScholarCross RefCross Ref
  16. [16] Ell Basil, Harth Andreas, and Simperl Elena. 2014. SPARQL query verbalization for explaining semantic search engine queries. In Proceedings of the European Semantic Web Conference. 426441.Google ScholarGoogle ScholarCross RefCross Ref
  17. [17] Fisher Aaron, Rudin Cynthia, and Dominici Francesca. 2018. Model class reliance: Variable importance measures for any machine learning model class, from the “Rashomon” perspective. arXiv preprint arXiv:1801.01489.Google ScholarGoogle Scholar
  18. [18] Friedman Jerome H.. 2001. Greedy function approximation: A gradient boosting machine. Annals of Statistics 29, 5 (2001), 11891232.Google ScholarGoogle ScholarCross RefCross Ref
  19. [19] Jerome H. Friedman and Bogdan E. Popescu. 2008. Predictive learning via rule ensembles. Annals of Applied Statistics 2, 3 (2008), 916954.Google ScholarGoogle ScholarCross RefCross Ref
  20. [20] Futia Giuseppe and Vetrò Antonio. 2020. On the integration of knowledge graphs into deep learning models for a more comprehensible AI—Three challenges for future research. Information 11, 2 (2020), 122.Google ScholarGoogle ScholarCross RefCross Ref
  21. [21] Gade Krishna, Geyik Sahin Cem, Kenthapadi Krishnaram, Mithal Varun, and Taly Ankur. 2019. Explainable AI in industry. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD’19). ACM, New York, NY, 32033204. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. [22] García María Vega and Aznarte José L.. 2020. Shapley additive explanations for NO2 forecasting. Ecological Informatics 56 (2020), 101039.Google ScholarGoogle ScholarCross RefCross Ref
  23. [23] Golbreich Christine, Wallace Evan K., and Patel-Schneider Peter F.. 2009. OWL 2 Web Ontology Language New Features and Rationale. W3C Working Draft 11 June 2009. Retrieved September 9, 2022 from http://www.w3.org/TR/2009/WD-owl2-new-features-20090611/.Google ScholarGoogle Scholar
  24. [24] Goldstein Alex, Kapelner Adam, Bleich Justin, and Pitkin Emil. 2015. Peeking inside the black box: Visualizing statistical learning with plots of individual conditional expectation. Journal of Computational and Graphical Statistics 24, 1 (2015), 4465.Google ScholarGoogle ScholarCross RefCross Ref
  25. [25] Greenwell Brandon M.. 2017. pdp: An R package for constructing partial dependence plots. R Journal 9, 1 (2017), 421436.Google ScholarGoogle ScholarCross RefCross Ref
  26. [26] Gunning David and Aha David W.. 2019. DARPA’s explainable artificial intelligence program. AI Magazine 40, 2 (2019), 4458.Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. [27] He Xin, Zhao Kaiyong, and Chu Xiaowen. 2019. AutoML: A survey of the state-of-the-art. arxiv:1908.00709 [cs.LG].Google ScholarGoogle Scholar
  28. [28] Hitzler Pascal, Bianchi Federico, Ebrahimi Monireh, and Sarker Kamruzzaman. 2020. Neural-symbolic integration and the semantic web. Semantic Web 11, 1 (2020), 311.Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. [29] Hitzler Pascal, Krötzsch Markus, Parsia Bijan, Patel-Schneider Peter F., and Rudolph Sebastian. 2009. OWL 2 web ontology language primer. W3C Recommendation 27, 1 (2009), 123.Google ScholarGoogle Scholar
  30. [30] Holzinger Andreas, Biemann Chris, Pattichis Constantinos S., and Kell Douglas B.. 2017. What do we need to build explainable AI systems for the medical domain? arXiv preprint arXiv:1712.09923.Google ScholarGoogle Scholar
  31. [31] Huang Yanping, Cheng Youlong, Bapna Ankur, Firat Orhan, Chen Dehao, Chen Mia, Lee HyoukJoong, et al. 2019. GPipe: Efficient training of giant neural networks using pipeline parallelism. Advances in Neural Information Processing Systems 32 (2019), 103112.Google ScholarGoogle Scholar
  32. [32] International SNOMED. 2022. Data Analytics with SNOMED CT. Retrieved June 29, 2022 from https://confluence.ihtsdotools.org/display/DOCANLYT/Data+Analytics+with+SNOMED+CT.Google ScholarGoogle Scholar
  33. [33] Bahador Khaleghi. 2020. An Explanation of What, Why, and How of eXplainable AI (XAI). Retrieved June 9, 2022 from https://towardsdatascience.com/an-explanation-of-what-why-and-how-of-explainable-ai-xai-117d9c441265.Google ScholarGoogle Scholar
  34. [34] Klaise Janis, Looveren Arnaud Van, Vacanti Giovanni, and Coca Alexandru. 2019. Alibi: Algorithms for Monitoring and Explaining Machine Learning Models. Retrieved September 9, 2022 from https://github.com/SeldonIO/alibi.Google ScholarGoogle Scholar
  35. [35] Kurakin Alexey, Goodfellow Ian J., and Bengio Samy. 2016. Adversarial machine learning at scale. CoRR abs/1611.01236.Google ScholarGoogle Scholar
  36. [36] Labs Accenture. 2020. Understanding Machines: Explainable AI. Technical Report. Retrieved September 9, 2022 from https://www.accenture.com/_acnmedia/pdf-85/accenture-understanding-machines-explainable-ai.pdf.Google ScholarGoogle Scholar
  37. [37] Li Mu, Andersen David G., Park Jun Woo, Smola Alexander J., Ahmed Amr, Josifovski Vanja, Long James, Shekita Eugene J., and Su Bor-Yiing. 2014. Scaling distributed machine learning with the parameter server. In Proceedings of the 11th USENIX Symposium on Operating Systems Design and Implementation (OSDI’14). 583–598.Google ScholarGoogle Scholar
  38. [38] Luciano Joanne S., Andersson Bosse, Batchelor Colin, Bodenreider Olivier, Clark Tim, Denney Christine K., Domarew Christopher, et al. 2011. The Translational Medicine Ontology and Knowledge Base: Driving personalized medicine by bridging the gap between bench and bedside. Journal of Biomedical Semantics 2 (2011), 121.Google ScholarGoogle Scholar
  39. [39] Lundberg Scott M., Erion Gabriel, Chen Hugh, DeGrave Alex, Prutkin Jordan M., Nair Bala, Katz Ronit, Himmelfarb Jonathan, Bansal Nisha, and Lee Su-In. 2020. From local explanations to global understanding with explainable AI for trees. Nature Machine Intelligence 2, 1 (2020), 25225839.Google ScholarGoogle ScholarCross RefCross Ref
  40. [40] Lundberg Scott M. and Lee Su-In. 2017. A unified approach to interpreting model predictions. In Advances in Neural Information Processing Systems. 47654774.Google ScholarGoogle Scholar
  41. [41] Mao Jiayuan, Gan Chuang, Kohli Pushmeet, Tenenbaum Joshua B., and Wu Jiajun. 2019. The neuro-symbolic concept learner: Interpreting scenes, words, and sentences from natural supervision. arXiv preprint arXiv:1904.12584.Google ScholarGoogle Scholar
  42. [42] Mirhoseini Azalia, Goldie Anna, Pham Hieu, Steiner Benoit, Le Quoc V., and Dean Jeff. 2018. A hierarchical model for device placement. In Proceedings of the International Conference on Learning Representations.Google ScholarGoogle Scholar
  43. [43] Mirhoseini Azalia, Pham Hieu, Le Quoc V., Steiner Benoit, Larsen Rasmus, Zhou Yuefeng, Kumar Naveen, Norouzi Mohammad, Bengio Samy, and Dean Jeff. 2017. Device placement optimization with reinforcement learning. In Proceedings of the International Conference on Machine Learning. 24302439.Google ScholarGoogle Scholar
  44. [44] Molnar Christoph. 2019. Interpretable Machine Learning. Retrieved September 9, 2022 from https://christophm.github.io/interpretable-ml-book/.Google ScholarGoogle Scholar
  45. [45] Mothilal Ramaravind Kommiya, Sharma Amit, and Tan Chenhao. 2019. Explaining machine learning classifiers through diverse counterfactual explanations. CoRR abs/1905.07697.Google ScholarGoogle Scholar
  46. [46] Ribeiro Marco Tulio, Singh Sameer, and Guestrin Carlos. 2018. Anchors: High-precision model-agnostic explanations. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence. 15271535.Google ScholarGoogle Scholar
  47. [47] Sarker Kamruzzaman, Xie Ning, Doran Derek, Raymer Michael L., and Hitzler Pascal. 2017. Explaining trained neural networks with semantic web technologies: First steps. In Proceedings of the Twelfth International Workshop on Neural-Symbolic Learning and Reasoning, NeSy 2017, London, UK, July 17-18, 2017(CEUR Workshop Proceedings, Vol. 2003), Besold Tarek R., Garcez Artur S. d’Avila, and Noble Isaac (Eds.). CEUR-WS.org. http://ceur-ws.org/Vol-2003/NeSy17_paper4.pdf.Google ScholarGoogle Scholar
  48. [48] Sarker Kamruzzaman, Zhou Lu, Eberhart Aaron, and Hitzler Pascal. 2021. Neuro-symbolic artificial intelligence: Current trends. arXiv preprint arXiv:2105.05330.Google ScholarGoogle Scholar
  49. [49] Seeliger Arne, Pfaff Matthias, and Krcmar Helmut. 2019. Semantic web technologies for explainable machine learning models: A literature review. PROFILES/SEMEX@ ISWC 2465 (2019), 116.Google ScholarGoogle Scholar
  50. [50] Selvaraju Ramprasaath R., Cogswell Michael, Das Abhishek, Vedantam Ramakrishna, Parikh Devi, and Batra Dhruv. 2017. Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision. 618626.Google ScholarGoogle ScholarCross RefCross Ref
  51. [51] Sergeev Alexander and Balso Mike Del. 2018. Horovod: Fast and easy distributed deep learning in TensorFlow. arXiv preprint arXiv:1802.05799.Google ScholarGoogle Scholar
  52. [52] Shrikumar Avanti, Greenside Peyton, and Kundaje Anshul. 2017. Learning important features through propagating activation differences. arXiv preprint arXiv:1704.02685.Google ScholarGoogle Scholar
  53. [53] Sigma Two. 2019. Interpretability Methods in Machine Learning: A Brief Survey. Retrieved September 9, 2022 from https://www.twosigma.com/insights/article/interpretability-methods-in-machine-learning-a-brief-survey/.Google ScholarGoogle Scholar
  54. [54] Smilkov Daniel, Thorat Nikhil, Kim Been, Viégas Fernanda, and Wattenberg Martin. 2017. SmoothGrad: Removing noise by adding noise. arXiv preprint arXiv:1706.03825.Google ScholarGoogle Scholar
  55. [55] Society The Royal. 2019. Explainable AI: The Basics, Policy Briefing. Retrieved September 9, 2022 from https://royalsociety.org/topics-policy/projects/explainable-ai/.Google ScholarGoogle Scholar
  56. [56] Steinruecken Christian, Smith Emma, Janz David, Lloyd James, and Ghahramani Zoubin. 2019. The automatic statistician. In Automated Machine Learning: Methods, Systems, Challenges, Hutter Frank, Kotthoff Lars, and Vanschoren Joaquin (Eds.). Springer, 161–164. Google ScholarGoogle ScholarCross RefCross Ref
  57. [57] Sundararajan Mukund, Taly Ankur, and Yan Qiqi. 2017. Axiomatic attribution for deep networks. arXiv preprint arXiv:1703.01365.Google ScholarGoogle Scholar
  58. [58] Thampi Ajay. 2020. Interpretable AI, Building Explainable Machine Learning Systems. Manning Publications. Google ScholarGoogle Scholar
  59. [59] Valiant Leslie G.. 2003. Three problems in computer science. Journal of the ACM 50, 1 (2003), 9699.Google ScholarGoogle ScholarDigital LibraryDigital Library
  60. [60] Wang Danding, Yang Qian, Abdul Ashraf, and Lim Brian Y.. 2019. Designing theory-driven user-centric explainable AI. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 115.Google ScholarGoogle ScholarDigital LibraryDigital Library
  61. [61] Xing Eric P., Ho Qirong, Dai Wei, Kim Jin Kyu, Wei Jinliang, Lee Seunghak, Zheng Xun, Xie Pengtao, Kumar Abhimanu, and Yu Yaoliang. 2015. Petuum: A new platform for distributed machine learning on big data. IEEE Transactions on Big Data 1, 2 (2015), 4967.Google ScholarGoogle ScholarCross RefCross Ref
  62. [62] Zhang Yongfeng and Chen Xu. 2018. Explainable recommendation: A survey and new perspectives. CoRR abs/1804.11192.Google ScholarGoogle Scholar
  63. [63] Vijay Arya, Rachel K. E. Bellamy, Pin-Yu Chen, Amit Dhurandhar, Michael Hind, Samuel C. Hoffman, Stephanie Houde, Q. Vera Liao, Ronny Luss, Aleksandra Mojsilović, Sami Mourad, Pablo Pedemonte, Ramya Raghavendra, John Richards, Prasanna Sattigeri, Karthikeyan Shanmugam, Moninder Singh, Kush R. Varshney, Dennis Wei, Yunfeng Zhang. 2019. One Explanation Does Not Fit All: A Toolkit and Taxonomy of AI Explainability Techniques. https://arxiv.org/abs/1909.03012.Google ScholarGoogle Scholar

Index Terms

  1. Explainable AI (XAI): Core Ideas, Techniques, and Solutions

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    • Published in

      cover image ACM Computing Surveys
      ACM Computing Surveys  Volume 55, Issue 9
      September 2023
      835 pages
      ISSN:0360-0300
      EISSN:1557-7341
      DOI:10.1145/3567474
      Issue’s Table of Contents

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 16 January 2023
      • Online AM: 4 September 2022
      • Accepted: 21 August 2022
      • Revised: 8 July 2022
      • Received: 20 October 2021
      Published in csur Volume 55, Issue 9

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • survey
      • Refereed

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    View Full Text

    HTML Format

    View this article in HTML Format .

    View HTML Format