Introduction
A Brief Historical Perspective
Objectives and Structure
-
This paper provides a completely revised hierarchical taxonomy that lays down the building block for future researchers to learn about the key aspects of XAI.
-
Using popular works in this field, we stress on the latest findings about XAI and its implications.
Fundamental Concepts of XAI
Black-Box Model
Interpretable Machine Learning
Types of Interpretability
-
Perceptive interpretability unifies all the interpretabilites that are well perceived by humans as they generally provide visual evidence. However, this obvious nature of the class lacks in fulfilling the true motive behind XAI because the black-box algorithm is yet to be unboxed. One of the integral methods to achieve perceptive interpretability is saliency which formulates its explanation based on the relative importance of all the input features. The resultant values could be in the form of probabilities (the LIME model [34]), super-pixels (ACE algorithm [40]), and heatmaps (CAM and LRP [41‐45]).
-
Interpretability by mathematical structures unifies all the interpretabilities that reveal the mechanisms behind deeper layers (which store all the complex information) of NN algorithms [46]. An example of this approach is testing with concept activation vectors (TCAV) [47]. Several other methods, such as t-distributed stochastic neighbor embedding (t-SNE) and correlation-based singular vector canonical correlation analysis (SVCCA) [48], play a significant role in directing towards the subspace of input for error-free predictions.
Abbreviation | Description |
---|---|
XAI | Explainable artificial intelligence |
FAT | Fairness-accountability-transparency |
ML | Machine learning |
DNN | Deep neural networks |
AGI | Artificial general intelligence |
MLP | Multilayer perceptrons |
IAI | Interpretable artificial intelligence |
DL | Deep learning |
CNN | Convolutional neural network |
RNN | Recurrent neural networks |
HCI | Human–computer interaction |
NLU | Natural language understanding |
ANI | Artificial narrow intelligence |
NLP | Natural language processing |
NN | Neural network |
Fundamental Concepts and Background
Interaction of Explainability With AI
-
Model interpretability: There is a growing focus on developing AI models that are interpretable, meaning that their decision-making process can be understood and explained to users.
-
Regulatory requirements: In some industries and regions, regulations are being introduced to require AI systems to be explainable in order to ensure accountability and transparency.
-
Trust and adoption: Explainability can play a role in building trust in AI systems, which is crucial for their widespread adoption and use.
-
Model validation: Explainability can help validate the decisions made by AI models, ensuring that they are free from biases and errors.
-
Debugging and improvement: Explainability can provide insights into how AI models are making decisions, making it easier to identify and address sources of error or bias.
Different Application Domains of XAI
Automated Transport
Medical
Financial
Related Works
References | Contributions | Strengths | Weaknesses |
---|---|---|---|
van der Velden et al. [73] | It contributes to improved trust and transparency in medical decision-making, allowing medical experts to understand why a specific diagnosis was made, leading to more accurate diagnoses and better patient outcomes. | Trends and future perspectives for XAI in medical image analysis are identified. | Struggles to cover all work in the field. |
Rudin et al. [74] | Identifies the top 10 challenges in interpretable ML along with their comprehensive background and solution approach. | Focused in assisting readers and new researchers to steer clear through common but problematic techniques related to interpretability in AI. | No discussion on socio-technical challenges, human–computer-interaction challenges and how robustness, as well as fairness, interact with interpretability. |
Abdul et al. [75] | The paper reveals fading and burgeoning trends in explainable systems and identifies closely related domains or mostly isolated. | Investigates how HCI researchers can help to develop accountable systems. | Lack of a range of philosophical theories about XAI and deep dive to extract information was lacking. |
Machlev et al. [76] | Highlights the potential of using XAI for power system applications. | This paper highlights the potential of using XAI for energy and power systems applications and covers the challenges and limitations of adopting and implementing XAI techniques in the field of energy and power systems. | The paper might favor specific industries or sectors and not take into account the needs and challenges faced by other industries using XAI in energy and power systems. |
Li et al. [3] | The progress in methodology, evaluation, and application of XAI is covered. | A new hierarchical taxonomy is explained which introduces the use of previous knowledge of XAI. | Overview of external knowledge is lacking and there are multiple open and unanswered questions. |
Tjoa et al. [39] | Aim to provide a comprehensive overview of the current state of XAI in the medical domain. | Aimed to give clinicians a perspective on the use of interpretable algorithms. | Less suitable for technically non-oriented readers due to some mathematical details. |
Adadi et al. [77] | Proposes the main concepts of enabling explainability in intelligent systems. | Includes a detailed discussion on the challenges and limitations of XAI methods, which allows for a more realistic understanding of the field and the potential areas of improvement. | It mainly focuses on the technical side of XAI and does not give enough attention to the ethical and societal implications of XAI. |
The Need For XAI
Why Explainability Matters
Need For Reasoning
Need For Innovation
Need For Regulation
Need For Advancement
Fair and Ethical Decision Making
-
It should comply with all laws and regulations and be lawful.
-
It should have good intentions and should be robust, both from a social and technical perspective.
-
It should demonstrate respect and should adhere to all principle and values.
-
For XAI to achieve this goal, it is important to carefully consider and address the key components of fair and ethical decision-making, including data quality, algorithmic bias, explainability, ethical principles, and human oversight.
-
Counterfactuals: This is a well-liked strategy for illuminating AI and making sense of algorithmic fairness. An excellent application of counterfactuals is risk management. A classic example of this is a bank. A bank may use counterfactual analysis to evaluate the potential outcomes of a loan application under different scenarios. For example, if the borrower loses their job or if interest rates increase unexpectedly. By simulating these potential scenarios, the bank can better understand the risks associated with the loan and make more informed decisions about whether or not to approve the application.
-
Fairness through awareness: This approach involves training AI models on data that is specifically designed to capture relevant factors contributing to fairness and ethical considerations. This can help to ensure that the models produce fair and ethical outcomes.
-
Fairness constraints: Fairness constraints are mathematical formulations of fairness criteria that can be used to optimise AI models. For example, a fairness constraint may be used to ensure that an AI model does not discriminate against certain demographic groups.
-
Human-in-the-loop approaches: Human-in-the-loop approaches involve incorporating human oversight into the decision-making process of AI systems. For example, a human may be involved in validating or adjusting the decisions made by an AI model to ensure that they are fair and ethical.
XAI Evaluation Frameworks
Design Principles For Interactive XAI
Adeptness to the User Behavior
Align Perception of XAI With Human Understandability
Collaborative Techniques Should Be Implemented More
Explanations Have More Dimensions Than Just Performance
Contradictions Provide Alternative Approaches
Are Explanations Always Important?
Amendments Go a Long Way
Techniques For XAI Implementation
Based on Scope
Global Interpretability
-
A desirable dataset X is chosen, which may either be the same black-box model’s training dataset or a brand-new dataset with comparable distributions.
-
The trained black-box model generates predictions from the chosen dataset.
-
To fit the black-box model’s predictions, an interpretable model is chosen and trained on dataset X.
Local Interpretability
-
From the predictions made by the black box, an instance of interest is chosen for which explanation is desired.
-
A fresh dataset is formed consisting of perturbed samples, and their respective predictions are extracted from the black-box model.
-
The new samples are weighted depending on their proximity to the instance of interest accordingly.
-
Now, the black-box model can be explained via an interpretable model trained on the perturbed dataset with help of the following equation [107]:$$\mathrm{Explanation}\;\left(x\right)\;=\;\arg\;\underset{g\epsilon G}{\min\;}L\;\left(f,\;g,\;\pi_x\right)\;+\;\Omega\left(g\right)$$(1)
Based on Stage
Ante-Hoc Interpretability
Post-Hoc Interpretability
Method | Advantages | Disadvantages |
---|---|---|
PDP | 1. Intuitive 2. Easy to implement 3. Interpretation is clear and causal | 1. Valid for maximum three features 2. Assumes absence of correlation between features 3. Heterogeneous effects may be hidden |
ICE | 1. Intuitive 2. Specific 3. Reveal heterogeneous relationships | 1. Overcrowding leads to unreadability 2. Need PDP to see the average |
ALE | 1. Unbiased towards correlated features 2. Faster Computation | 1. Unsteady with high number of intervals 2. Comparatively much more complex 3. Not accompanied by ICE plots |
LIME | 1. Inherently interpretable 2. Widely acceptable especially in DNN’s 3. Human-Freindly explanations | 1. Unsatisfactory global approximation 2. Easily manipulated to hide biases |
-
Interactive visualisations: Using dynamic, interactive visualisations to allow users to explore data and model behavior in real-time.
-
Augmented Reality (AR) and Virtual Reality (VR) visualisations: Using AR and VR technology to create immersive visualisations that can help users better understand and interact with data.
-
3D and 4D visualisations: Using 3D and 4D visualisations to represent data in new and more informative ways, allowing users to better understand complex relationships and patterns.
-
Automated visualisation generation: Using machine learning algorithms to automatically generate visualisations based on the data and desired output.
-
Real-time streaming visualisations: Creating visualisations that can be updated in real-time as new data is received, allowing users to monitor the data and model performance in real-time.
-
Multi-view visualisations: Using multiple views or perspectives to represent data, allowing users to explore and understand the data from different angles.
-
Application grounded—This methodology deals with domain experts experimenting in real-life situations to validate the delivery of end-task. More specifically, it acknowledges any discovery of new facts, debugging of new errors and elimination of biases. For example, to diagnose a disease, the most suitable way is for a doctor to perform diagnosis [160].
-
Human grounded—This methodology deals with lay humans performing general experiments to address a wider pool of evaluation concepts. This proves to be a cost-effective method of maintaining the crux of the target application. For example, humans are given a choice between two theories, and they need to choose the one with higher accuracy [161].
-
Functionally Grounded—This methodology deals with a proxy measure for evaluating interpretability, and therefore additional research is needed. No requirement for human interaction and minimal cost are some appealing factors for its wide usage. For example, decision trees are considered interpretable in many situations [162].
Challenges For Enabling XAI
Human-Machine Collaboration
-
Improved decision quality: By leveraging the strengths of both humans and machines, human–machine collaboration can result in improved decision quality compared to relying on either alone. This idea emerges because of the compulsive nature to improve the reliability of the system [173].
-
Increased transparency: Human–machine collaboration can increase transparency and accountability in AI systems, as humans can understand the reasoning behind the decisions made by these systems. This will allow shared awareness and intent for optimal teamwork [174].
-
Mitigation of bias: By incorporating human feedback and oversight, human–machine collaboration can help to mitigate bias and ensure that AI systems are fair and ethical. Using an end-to-end machine learning pipeline which includes pre-processing, in-processing, and post-processing, data scientists can detect and eliminate any form of bias in their models.
-
Better performance over time: Human–machine collaboration allows AI systems to continuously improve over time, as human experts can provide feedback and guidance to improve the performance of these systems.
Acceptance and Popularisation
Discussion and Reflection
Future Research Directions
Responsible AI
-
Acceptance of responsibility will determine public attitude towards the acceptance of responsible AI in society. Governments and citizens need to act together to resolve issues of reliability concerned with AI.
-
Self-justification will enable AI models to develop reasoning and a code of conduct based on human values and ethics. Current research shows that an adequate link between decisions and ethical context is missing.
-
Participation involves the real-life application of AI in everyday life to develop the guidelines for responsible AI. Here, education plays a significant role in creating awareness among people that their input is crucial in shaping the societal character of responsible AI.