Keywords: Deep Learning Interpretability Explainable AI
Over the years, ML models have steadily grown in complexity, gaining predictivity often at the expense of interpretability. An active research area called explainable AI (or XAI) has emerged with the goal to produce models that are both predictive and understandable. XAI has reached important successes, such as robust heatmap-based explanations of DNN classifiers. From an application perspective, there is now a need to massively engage into new scenarios such as explaining unsupervised / reinforcement learning, as well as producing explanations that are optimally structured for the human. In particular, our planned workshop will cover the following topics:
- Explaining beyond DNN classifiers: random forests, unsupervised learning, reinforcement learning
- Explaining beyond heatmaps: structured explanations, Q/A and dialog systems, human-in-the-loop
- Explaining beyond explaining: Improving ML models and algorithms, verifying ML, getting insights
XAI has received an exponential interest in the research community, and awareness of the need to explain ML models have grown in similar proportions in industry and in the sciences. With the sizable XAI research community that has formed, there is now a key opportunity to achieve this push towards successful applications. Our hope is that our proposed XXAI workshop can accelerate this process, foster a more systematic use of XAI to produce improvement on models in applications, and finally, also serves to better identify in which way current XAI methods need to be improved and what kind of theory of XAI is needed.
Thu 11:45 p.m. - 12:00 a.m.
|
Opening Remarks
(
Talk
)
|
Wojciech Samek 🔗 |
Fri 12:00 a.m. - 12:25 a.m.
|
Invited Talk 1: Scott Lundberg - From local explanations to global understanding with trees
(
Talk
)
link »
SlidesLive Video » Tree-based machine learning models are popular nonlinear predictive models, yet comparatively little attention has been paid to explaining their predictions. In this talk I will explain how to improve their interpretability through the combination of many local game-theoretic explanations. I'll show how combining many high-quality local explanations allows us to represent global structure while retaining local faithfulness to the original model. This will enable us to identify high-magnitude but low-frequency nonlinear mortality risk factors in the US population, to highlight distinct population subgroups with shared risk characteristics, and to identify nonlinear interaction effects among risk factors for chronic kidney disease, and to monitor a machine learning model deployed in a hospital by identifying which features are degrading the model’s performance over time. |
Wojciech Samek 🔗 |
Fri 12:25 a.m. - 12:30 a.m.
|
Invited Talk 1 Q&A
(
Q&A
)
|
🔗 |
Fri 12:30 a.m. - 12:55 a.m.
|
Invited Talk 2: Bolei Zhou - Interpreting and Leveraging the Latent Semantics in Deep Generative Models
(
Talk
)
link »
SlidesLive Video » Recent progress in deep generative models such as Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs) has enabled synthesizing photo-realistic images, such as faces and scenes. However, it remains much less explored on what has been learned inside the deep representations learned from synthesizing images. In this talk, I will present some of our recent progress in interpreting the semantics in the latent space of the GANs, as well as reversing real images back into the latent space. Identifying these semantics not only allows us to better understand the internal mechanism in generative models, but also facilitates versatile real image editing applications. |
Wojciech Samek 🔗 |
Fri 12:55 a.m. - 1:00 a.m.
|
Invited Talk 2 Q&A
(
Q&A
)
|
🔗 |
Fri 1:00 a.m. - 1:15 a.m.
|
Contributed Talk 1: Sun et al. - Understanding Image Captioning Models beyond Visualizing Attention
(
Talk
)
link »
This paper explains predictions of image captioning attention models beyond visualizing the attention itself. In this paper, we develop variants of layer-wise relevance backpropagation (LRP) tailored to image captioning models with attention mechanisms. We show that the explanations, firstly, correlate to object locations with higher precision than attention, secondly, identify object words that are unsupported by image content, and thirdly, provide guidance to improve the model. Results are reported using two different image captioning attention models trained with Flickr30K and MSCOCO2017 datasets. Experimental analyses show the strength of explanation methods for understanding image captioning attention models. |
Wojciech Samek 🔗 |
Fri 1:15 a.m. - 1:30 a.m.
|
Contributed Talk 2: Karimi et al. - Algorithmic recourse under imperfect causal knowledge: a probabilistic approach
(
Tak
)
link »
SlidesLive Video » Recent work has discussed the limitations of counterfactual explanations to recommend actions for algorithmic recourse and argued for the need of taking causal relationships between features into consideration. Unfortunately, in practice, the true underlying structural causal model is generally unknown. In this work, we first show that it is impossible to guarantee recourse without access to the true structural equations. To address this limitation, we propose two probabilistic approaches to select optimal actions that achieve recourse with high probability given limited causal knowledge (e.g., only the causal graph). The first captures uncertainty over structural equations under additive Gaussian noise and uses Bayesian model averaging to estimate the counterfactual distribution. The second removes any assumptions on the structural equations by instead computing the average effect of recourse actions on individuals similar to the person who seeks recourse, leading to a novel subpopulation-based interventional notion of recourse. We derive a gradient-based procedure for selecting optimal recourse actions and empirically show that the proposed approaches lead to more reliable recommendations under imperfect causal knowledge than non-probabilistic baselines. |
Wojciech Samek 🔗 |
Fri 1:30 a.m. - 3:00 a.m.
|
Poster Session 1
(
Parallel Poster Sessions
)
Zoom Room 1
[ protected link dropped ] Zoom Room 2
[ protected link dropped ] |
Wojciech Samek 🔗 |
Fri 3:00 a.m. - 3:25 a.m.
|
Invited Talk 3: Grégoire Montavon - XAI Beyond Classifiers: Explaining Anomalies, Clustering, and More
(
Talk
)
link »
SlidesLive Video » Unsupervised models such as clustering or anomaly detection are routinely used for data discovery and summarization. To gain maximum insight from the data, we also need to explain which input features (e.g. pixels) support the cluster assignments and the anomaly detections.—So far, XAI has mainly focused on supervised models.—In this talk, a novel systematic approach to explain various unsupervised models is presented. The approach is based on finding, without retraining, neural network equivalents of these models. Their predictions can then be readily explained using common XAI procedures developed for neural networks. |
Wojciech Samek 🔗 |
Fri 3:25 a.m. - 3:30 a.m.
|
Invited Talk 3 Q&A
(
Q&A
)
|
🔗 |
Fri 3:30 a.m. - 3:55 a.m.
|
Invited Talk 4: Zeynep Akata - Modelling Conceptual Understanding Through Communication
(
Talk
)
link »
SlidesLive Video » An agent who interacts with a wide population of other agents needs to be aware that there may be variations in their understanding of the world. Furthermore, the machinery which they use to perceive may be inherently different, as is the case between humans and machines. In this work, we present both an image reference game between a speaker and a population of listeners where reasoning about the concepts other agents can comprehend is necessary and a model formulation with this capability. We focus on reasoning about the conceptual understanding of others, as well as adapting to novel gameplay partners and dealing with differences in perceptual machinery. Our experiments on three benchmark image/attribute datasets suggest that our learner indeed encodes information directly pertaining to the understanding of other agents, and that leveraging this information is crucial for maximizing gameplay performance. |
Wojciech Samek 🔗 |
Fri 3:55 a.m. - 4:00 a.m.
|
Invited Talk 4 Q&A
(
Q&A
)
|
🔗 |
Fri 4:00 a.m. - 4:25 a.m.
|
Invited Talk 5: Sepp Hochreiter - XAI and Strategy Extraction via Reward Redistribution
(
Talk
)
link »
SlidesLive Video » Assigning credit for a received reward to previously performed actions is one of the central tasks in reinforcement learning. Credit assignment often uses world models, either in a forward or in a backward view. In a forward view, the future return is estimated by replacing the environment through a model or by rolling out sequences until episode end. A backward view either learns a backward model or performs a backward analysis of a forward model that predicts or models the return of an episode. Our method RUDDER performs a backward analysis to construct a reward redistribution to credit those actions that caused a reward. Its extension Align-RUDDER learns a reward redistribution from few demonstrations. An optimal reward redistribution has zero expected future reward and, therefore, immediately credits actions for all they will cause. XAI aims at credit assignment, too, when asking what caused a model to produce a particular output given an input. Even further, XAI wants to know how and why a policy solved a task, why an agent is better than humans, why a decision was made. Humans best comprehend a strategy of an agent if all its actions are immediately evaluated and do not have hidden consequences in the future. Reward redistributions learned by RUDDER and Align-RUDDER help to understand task-solving strategies of both humans and machines. |
Wojciech Samek 🔗 |
Fri 4:25 a.m. - 4:30 a.m.
|
Invited Talk 5 Q&A
(
Q&A
)
|
🔗 |
Fri 4:30 a.m. - 4:55 a.m.
|
Invited Talk 6: Ribana Roscher - Use of Explainable Machine Learning in the Sciences
(
Talk
)
link »
SlidesLive Video » For some time now, machine learning methods have been indispensable in many application areas. Especially with the recent development of neural networks, these methods are increasingly used in the sciences to obtain scientific outcomes from observational or simulated data. Besides a high accuracy, a desired goal is to learn explainable models. In order to reach this goal and obtain explanation, knowledge from the respective domain is necessary, which can be integrated into the model or applied post-hoc. This talk focuses on explainable machine learning approaches which are used to tackle common challenges in the sciences such as the provision of reliable and scientific consistent results. It will show that recent advances in machine learning to enhance transparency, interpretability, and explainability are helpful in overcoming these challenges. |
Wojciech Samek 🔗 |
Fri 4:55 a.m. - 5:00 a.m.
|
Invited Talk 6 Q&A
(
Q&A
)
|
🔗 |
Fri 5:00 a.m. - 5:25 a.m.
|
Invited Talk 7: Adrian Weller & Umang Bhatt - Challenges in Deploying Explainable Machine Learning
(
Talk
)
link »
SlidesLive Video » Explainable machine learning offers the potential to provide stakeholders with insights into model behavior, yet there is little understanding of how organizations use these methods in practice. In this talk, we discuss recent research exploring how organizations view and use explainability. We find that the majority of deployments are not for end-users but rather for machine learning engineers, who use explainability to debug the model. There is thus a gap between explainability in practice and the goal of external transparency since explanations are primarily serving internal stakeholders. Providing useful external explanations requires careful consideration of the needs of stakeholders, including end-users, regulators, and domain experts. Despite this need, little work has been done to facilitate inter-stakeholder conversation around explainable machine learning. To help address this gap, we report findings from a closed-door, day-long workshop between academics, industry experts, legal scholars, and policymakers to develop a shared language around explainability and to understand the current shortcomings of and potential solutions for deploying explainable machine learning in the service of external transparency goals. |
Wojciech Samek 🔗 |
Fri 5:25 a.m. - 5:30 a.m.
|
Invited Talk 7 Q&A
(
Q&A
)
|
🔗 |
Fri 5:30 a.m. - 5:55 a.m.
|
Invited Talk 8: Osbert Bastani - Interpretable, Robust, and Verifiable Reinforcement Learning
(
Talk
)
link »
SlidesLive Video » Structured control policies such as decision trees, finite-state machines, and programs have a number of advantages over more traditional models: they are easier for humans to understand and debug, they generalize more robustly to novel environments, and they are easier to formally verify. However, learning these kinds of models has proven to be challenging. I will describe recent progress learning structured policies, along with evidence demonstrating their benefits. |
Wojciech Samek 🔗 |
Fri 5:55 a.m. - 6:00 a.m.
|
Invited Talk 8 Q&A
(
Q&A
)
|
🔗 |
Fri 6:00 a.m. - 6:15 a.m.
|
Contributed Talk 3: Anders et al. - XAI for Analyzing and Unlearning Spurious Correlations in ImageNet
(
Talk
)
link »
SlidesLive Video » Contemporary learning models for computer vision are typically trained on very large data sets with millions of samples. There may, however, be biases, artifacts, or errors in the data that have gone unnoticed and are exploitable by the model, which in turn becomes a biased ‘Clever-Hans‘ predictor. In this paper, we contribute by providing a comprehensive analysis framework based on a scalable statistical analysis of attributions from explanation methods for large data corpora, here ImageNet. Based on Spectral Relevance Analysis we propose the following technical contributions and resulting findings: (a) a scalable quantification of artifactual classes where the ML models under study exhibit Clever-Hans behavior, (b) an approach denoted as Class-Artifact Compensation (ClArC) that allows to fine-tune an existing model to effectively eliminate its focus on artifacts and biases yielding significantly reduced Clever-Hans behavior. |
Wojciech Samek 🔗 |
Fri 6:15 a.m. - 6:30 a.m.
|
Contributed Talk 4: Yau et al. - What did you think would happen? Explaining Agent Behaviour through Intended Outcomes
(
Talk
)
link »
SlidesLive Video » We present a novel form of explanation for Reinforcement Learning (RL), based around the notion of intended outcome. This describes what outcome an agent is trying to achieve by its actions. Given this definition, we provide a simple proof that general methods for post-hoc explanations of this nature are impossible in traditional reinforcement learning. Rather, the information needed for the explanations must be collected in conjunction with training the agent. We provide approaches designed to do this for several variants of Q-function approximation and prove consistency between the explanations and the Q-values learned. We demonstrate our method on multiple reinforcement learning problems. |
Wojciech Samek 🔗 |
Fri 6:30 a.m. - 8:00 a.m.
|
Poster Session 2
(
Parallel Poster Sessions
)
Zoom Room 1
[ protected link dropped ] Zoom Room 2
[ protected link dropped ] Zoom Room 3
[ protected link dropped ] Zoom Room 4
[ protected link dropped ] Zoom Room 5
[ protected link dropped ] |
Wojciech Samek 🔗 |
Fri 8:00 a.m. - 8:10 a.m.
|
Closing Remarks
(
Discussion
)
|
Wojciech Samek 🔗 |