Workshop: Responsible Decision Making in Dynamic Environments

A law of adversarial risk, interpolation, and label noise

Daniel Paleka · Amartya Sanyal


In supervised learning, it is known that label noise in the data can be interpolated without penalties on test accuracy. We show that interpolating label noise induces adversarial vulnerability, and prove the first theorem showing the dependence of label noise and adversarial risk in terms of the data distribution. Our results are almost sharp without accounting for the inductive bias of the learning algorithm. We also show that inductive bias makes the effect of label noise much stronger.

Chat is not available.