Skip to yearly menu bar Skip to main content


Poster

SPADE: Sparsity-Guided Debugging for Deep Neural Networks

Arshia Soltani Moakhar · Eugenia Iofinova · Elias Frantar · Dan Alistarh


Abstract:

It is known that sparsity can improve interpretability for deep neural networks. However, existing methods in the area either require networks that are pre-trained with sparsity constraints, or impose sparsity after the fact, altering the network's general behavior. In this paper, we demonstrate, for the first time, that sparsity can instead be incorporated into the interpretation process itself, as a sample-specific preprocessing. Unlike previous work, this approach, which we call SPADE, does not place constraints on the trained model and does not affect its behavior during inference on the sample. Given a trained model and a target sample, SPADE uses sample-targeted pruning to provide a "trace" of the network's execution on the sample, reducing the network to the connections that are most relevant to the specific prediction. We demonstrate that preprocessing with SPADE significantly increases the accuracy of image saliency maps across several interpretability methods. Additionally, SPADE improves the usefulness of neuron visualizations, aiding humans in reasoning about network behavior. We demonstrate that the latter effect is due to the fact that pruning can disentangle polysemantic neurons, consistently leading to improved interpretations.

Live content is unavailable. Log in and register to view live content