Skip to yearly menu bar Skip to main content


( events)   Timezone:  
Workshop
Wed Aug 09 03:30 PM -- 12:30 AM (PDT) @ C4.8
Workshop on Human Interpretability in Machine Learning (WHI)
Kush Varshney · Adrian Weller · Been Kim · Dmitry Malioutov





Workshop Home Page

This workshop will bring together researchers who study the interpretability of predictive models, develop interpretable machine learning algorithms, and develop methodology to interpret black-box machine learning models (e.g., post-hoc interpretations). This is a very exciting time to study interpretable machine learning, as the advances in large-scale optimization and Bayesian inference that have enabled the rise of black-box machine learning are now also starting to be exploited to develop principled approaches to large-scale interpretable machine learning. Participants in the workshop will exchange ideas on these and allied topics, including:

● Quantifying and axiomatizing interpretability
● Psychology of human concept learning
● Rule learning,Symbolic regression and case-based reasoning
● Generalized additive models, sparsity and interpretability
● Visual analytics
● Interpretable unsupervised models (clustering, topic models, e.t.c)
● Interpretation of black-box models (including deep neural networks)
● Causality of predictive models
● Verifying, diagnosing and debugging machine learning systems
● Interpretability in reinforcement learning.

Doctors, judges, business executives, and many other people are faced with making critical decisions that can have profound consequences. For example, doctors decide which treatment to administer to patients, judges decide on prison sentences for convicts, and business executives decide to enter new markets and acquire other companies. Such decisions are increasingly being supported by predictive models learned by algorithms from historical data.

The latest trend in machine learning is to use highly nonlinear complex systems such as deep neural networks, kernel methods, and large ensembles of diverse classifiers. While such approaches often produce impressive, state-of-the art prediction accuracies, their black-box nature offers little comfort to decision makers. Therefore, in order for predictions to be adopted, trusted, and safely used by decision makers in mission-critical applications, it is imperative to develop machine learning methods that produce interpretable models with excellent predictive accuracy. It is in this way that machine learning methods can have impact on consequential real-world applications.

A. Dhurandhar, V. Iyengar, R. Luss, and K. Shanmugam, "A Formal Framework to Characterize Interpretability of Procedures" (Contributed Talk)
A. Henelius, K. Puolamäki, and A. Ukkonen, "Interpreting Classifiers through Attribute Interactions in Datasets" (Contributed Talk)
S. Lundberg and S.-I. Lee, "Consistent Feature Attribution for Tree Ensembles" (Contributed Talk)
Invited Talk: D. Sontag (Invited Talk)
S. Penkov and S. Ramamoorthy, "Program Induction to Interpret Transition Systems" (Contributed Talk)
R. L. Phillips, K. H. Chang, and S. Friedler, "Interpretable Active Learning" (Contributed Talk)
C. Rosenbaum, T. Gao, and T. Klinger, "e-QRAQ: A Multi-turn Reasoning Dataset and Simulator with Explanations" (Contributed Talk)
Invited Talk: T. Jebara, "Interpretability Through Causality" (Invited Talk)
W. Tansey, J. Thomason, and J. G. Scott, "Interpretable Low-Dimensional Regression via Data-Adaptive Smoothing" (Contributed Talk)
Invited Talk: P. W. Koh (Invited Talk)
I. Valera, M. F. Pradier, and Z. Ghahramani, "General Latent Feature Modeling for Data Exploration Tasks" (Contributed Talk)
A. Weller, "Challenges for Transparency" (Contributed Talk)
ICML WHI 2017 Awards Ceremony (Awards Ceremony)
Panel Discussion: Human Interpretability in Machine Learning (Panel Discussion)