The Mysteries of Adversarial Robustness for Non-parametric Methods and Neural Networks
Speaker: Kamalika Chaudhuri (University of California, San Diego)
Occasion: The Multifaceted Complexity of Machine Learning
Date: April 16, 2021
Abstract: Adversarial examples are small imperceptible perturbations to legitimate test inputs that cause machine learning classifiers to misclassify. While recent work has proposed many attacks and defenses, why exactly they arise still remains a mystery. In this talk, we’ll take a closer look at this question.
We will look at non-parametric methods, and define a large sample limit for adversarially robust classification that is analogous to the Bayes optimal. We will show then that adversarial robustness in non-parametric methods is mostly a consequence of the training method. If time permits, we will then look at what these findings mean for neural networks.