Skip to yearly menu bar Skip to main content


Poster
in
Workshop: 2nd Annual Workshop on Topology, Algebra, and Geometry in Machine Learning (TAG-ML)

On the Relationship Between Data Manifolds and Adversarial Examples

Michael Geyer · Brian Bell · Amanda Fernandez · Juston Moore


Abstract:

In this work we study adversarial examples in deep neural networks through the lens of a predefined data manifold.By forcing certain geometric properties of this manifold, we are able to analyze the behavior of the learned decision boundaries.It has been shown previously that training to be robust against adversarial attacks produces models with gradients aligned to a small set of principal variations in the data. We demonstrate the converse of this statement; aligning model gradients with a select set of principal variations improves robustness against gradient based adversarial attacks. Our analysis shows that this also makes data more orthogonal to decision boundaries. We conclude that robust training methods make the problem better posed by focusing the model on more important dimensions of variation.

Chat is not available.