Timezone: »

A law of adversarial risk, interpolation, and label noise
Daniel Paleka · Amartya Sanyal

In supervised learning, it is known that label noise in the data can be interpolated without penalties on test accuracy. We show that interpolating label noise induces adversarial vulnerability, and prove the first theorem showing the dependence of label noise and adversarial risk in terms of the data distribution. Our results are almost sharp without accounting for the inductive bias of the learning algorithm. We also show that inductive bias makes the effect of label noise much stronger.

Author Information

Daniel Paleka (ETH Zurich)
Amartya Sanyal (University of Oxford)

More from the Same Authors