Timezone: »
On the Connections between Counterfactual Explanations and Adversarial Examples
· Martin Pawelczyk · Shalmali Joshi · Chirag Agarwal · Sohini Upadhyay · Hima Lakkaraju
Author Information
Martin Pawelczyk (University of Tuebingen)
Shalmali Joshi (Harvard University (SEAS))
Chirag Agarwal (Harvard University)
Sohini Upadhyay (Harvard University)
Hima Lakkaraju (Harvard)
More from the Same Authors
-
2021 : Poster Session Test »
Jie Ren · -
2021 : A Turing Test for Transparency »
· Felix Biessmann -
2021 : Dynamic Defense Approach for Adversarial Robustness in Deep Neural Networks via Stochastic Ensemble Smoothed Model »
· Ruoxi Qin -
2021 : Quality Metrics for Transparent Machine Learning With and Without Humans In the Loop Are Not Correlated »
· Felix Biessmann -
2021 : Minimum sharpness: Scale-invariant parameter-robustness of neural networks »
· Hikaru Ibayashi -
2021 : Understanding Instance-based Interpretability of Variational Auto-Encoders »
· Zhifeng Kong · Kamalika Chaudhuri -
2021 : Informative Class Activation Maps: Estimating Mutual Information Between Regions and Labels »
· Zhenyue Qin -
2021 : This Looks Like That... Does it? Shortcomings of Latent Space Prototype Interpretability in Deep Networks »
· Adrian Hoffmann · Claudio Fanconi · Rahul Rade · Jonas Kohler -
2021 : How Not to Measure Disentanglement »
· Julia Kiseleva · Maarten de Rijke -
2021 : Diverse and Amortised Counterfactual Explanations for Uncertainty Estimates »
· Dan Ley · Umang Bhatt · Adrian Weller -
2021 : Towards the Unification and Robustness of Perturbation and Gradient Based Explanations »
· Sushant Agarwal · Shahin Jabbari · Chirag Agarwal · Sohini Upadhyay · Steven Wu · Hima Lakkaraju -
2021 : Improving the Transferability of Adversarial Examples with New Iteration Framework and Input Dropout »
· Pengfei Xie -
2021 : Interpretable Face Manipulation Detection via Feature Whitening »
· Yingying Hua · Pengju Wang · Shiming Ge -
2021 : Synthetic Benchmarks for Scientific Research in Explainable Machine Learning »
· Yang Liu · Colin White · Willie Neiswanger -
2021 : A Probabilistic Representation of DNNs: Bridging Mutual Information and Generalization »
· Xinjie Lan -
2021 : A MaxSAT Approach to Inferring Explainable Temporal Properties »
· Rajarshi Roy · Zhe Xu · Ufuk Topcu · Jean-RaphaĆ«l Gaglione -
2021 : Active Automaton Inference for Reinforcement Learning using Queries and Counterexamples »
· Aditya Ojha · Zhe Xu · Ufuk Topcu -
2021 : Learned Interpretable Residual Extragradient ISTA for Sparse Coding »
· Connie Kong · Fanhua Shang -
2021 : Neural Network Classifier as Mutual Information Evaluator »
· Zhenyue Qin -
2021 : Evaluation of Saliency-based Explainability Methods »
· Sam Zabdiel Samuel · Vidhya Kamakshi · Narayanan Chatapuram Krishnan -
2021 : Order in the Court: Explainable AI Methods Prone to Disagreement »
· Michael Neely · Stefan F. Schouten · Ana Lucic -
2021 : On the overlooked issue of defining explanation objectives for local-surrogate explainers »
· Rafael Poyiadzi · Xavier Renard · Thibault Laugel · Raul Santos-Rodriguez · Marcin Detyniecki -
2021 : How Well do Feature Visualizations Support Causal Understanding of CNN Activations? »
· Roland S. Zimmermann · Judith Borowski · Robert Geirhos · Matthias Bethge · Thomas SA Wallis · Wieland Brendel -
2021 : Promises and Pitfalls of Black-Box Concept Learning Models »
· Anita Mahinpei · Justin Clark · Isaac Lage · Finale Doshi-Velez · Weiwei Pan -
2021 : On the (Un-)Avoidability of Adversarial Examples »
· Ruth Urner -
2021 : Towards a Rigorous Theoretical Analysis and Evaluation of GNN Explanations »
· Chirag Agarwal · Marinka Zitnik · Hima Lakkaraju -
2021 : Reliable graph neural network explanations through adversarial training »
· Donald Loveland · Bhavya Kailkhura · T. Yong-Jin Han -
2021 : Towards Fully Interpretable Deep Neural Networks: Are We There Yet? »
· Sandareka Wickramanayake -
2021 : Towards Automated Evaluation of Explanations in Graph Neural Networks »
· Balaji Ganesan · Devbrat Sharma -
2021 : A Source-Criticism Debiasing Method for GloVe Embeddings »
-
2021 : Quantitative Evaluation of Explainable Graph Neural Networks for Molecular Property prediction »
· Jiahua Rao · SHUANGJIA ZHENG -
2021 : What will it take to generate fairness-preserving explanations? »
· Jessica Dai · Sohini Upadhyay · Hima Lakkaraju -
2021 : Gradient-Based Interpretability Methods and Binarized Neural Networks »
· Amy Widdicombe -
2021 : Meaningfully Explaining a Model's Mistakes »
· Abubakar Abid · James Zou -
2021 : Feature Attributions and Counterfactual Explanations Can Be Manipulated »
· Dylan Slack · Sophie Hilgard · Sameer Singh · Hima Lakkaraju -
2021 : SalKG: Learning From Knowledge Graph Explanations for Commonsense Reasoning »
· Aaron Chan · Xiang Ren -
2021 : Re-imagining GNN Explanations with ideas from Tabular Data »
· Anjali Singh · Shamanth Nayak K · Balaji Ganesan -
2021 : Learning Sparse Representations with Alternating Back-Propagation »
· Tian Han -
2021 : Deep Interpretable Criminal Charge Prediction Based on Temporal Trajectory »
· Jia Xu · Abdul Khan -
2021 : On the Connections between Counterfactual Explanations and Adversarial Examples »
Martin Pawelczyk · Shalmali Joshi · Chirag Agarwal · Sohini Upadhyay · Hima Lakkaraju -
2021 : Towards Robust and Reliable Algorithmic Recourse »
Sohini Upadhyay · Shalmali Joshi · Hima Lakkaraju -
2021 : CARLA: A Python Library to Benchmark Algorithmic Recourse and Counterfactual Explanation Algorithms »
Martin Pawelczyk · Sascha Bielawski · Gjergji Kasneci -
2021 : Enhancing interpretability and reducing uncertainties in deep learning of electrocardiograms using a sub-waveform representation »
Hossein Honarvar · Chirag Agarwal · Sulaiman Somani · Girish Nadkarni · Marinka Zitnik · Fei Wang · Benjamin Glicksberg -
2021 : Reliable Post hoc Explanations: Modeling Uncertainty in Explainability »
Dylan Slack · Sophie Hilgard · Sameer Singh · Hima Lakkaraju -
2021 : Interpretable learning-to-defer for sequential decision-making »
Shalmali Joshi · Sonali Parbhoo · Finale Doshi-Velez -
2021 : Towards a Unified Framework for Fair and Stable Graph Representation Learning »
Chirag Agarwal · Hima Lakkaraju · Marinka Zitnik -
2021 : Reliable Post hoc Explanations: Modeling Uncertainty in Explainability »
Dylan Slack · Sophie Hilgard · Sameer Singh · Hima Lakkaraju -
2021 : Interpretable learning-to-defer for sequential decision-making »
Shalmali Joshi · Sonali Parbhoo · Finale Doshi-Velez -
2021 : On formalizing causal off-policy sequential decision-making »
Sonali Parbhoo · Shalmali Joshi · Finale Doshi-Velez -
2022 : "Why did the Model Fail?": Attributing Model Performance Changes to Distribution Shifts »
Haoran Zhang · Harvineet Singh · Shalmali Joshi -
2023 Tutorial: Responsible AI for Generative AI in Practice: Lessons Learned and Open Challenges »
Hima Lakkaraju · Krishnaram Kenthapadi · Nazneen Rajani -
2022 Workshop: New Frontiers in Adversarial Machine Learning »
Sijia Liu · Pin-Yu Chen · Dongxiao Zhu · Eric Wong · Kathrin Grosse · Hima Lakkaraju · Sanmi Koyejo -
2021 Workshop: ICML Workshop on Algorithmic Recourse »
Stratis Tsirtsis · Amir-Hossein Karimi · Ana Lucic · Manuel Gomez-Rodriguez · Isabel Valera · Hima Lakkaraju -
2021 : Invited Speakers' Panel »
Neeraja J Yadwadkar · Shalmali Joshi · Roberto Bondesan · Engineer Bainomugisha · Stephen Roberts -
2021 : Towards Robust and Reliable Model Explanations for Healthcare »
Hima Lakkaraju -
2021 : Ethics of developing ML in healthcare »
Shalmali Joshi -
2021 Poster: Towards the Unification and Robustness of Perturbation and Gradient Based Explanations »
Sushant Agarwal · Shahin Jabbari · Chirag Agarwal · Sohini Upadhyay · Steven Wu · Hima Lakkaraju -
2021 Spotlight: Towards the Unification and Robustness of Perturbation and Gradient Based Explanations »
Sushant Agarwal · Shahin Jabbari · Chirag Agarwal · Sohini Upadhyay · Steven Wu · Hima Lakkaraju -
2020 Poster: Robust and Stable Black Box Explanations »
Hima Lakkaraju · Nino Arsov · Osbert Bastani