Skip to yearly menu bar Skip to main content


Spotlight Poster

On the Role of Randomization in Adversarially Robust Classification

Lucas Gnecco Heredia · Muni Sreenivas Pydi · Laurent Meunier · Benjamin Negrevergne · Yann Chevaleyre

Great Hall & Hall B1+B2 (level 1) #806

Abstract:

Deep neural networks are known to be vulnerable to small adversarial perturbations in test data. To defend against adversarial attacks, probabilistic classifiers have been proposed as an alternative to deterministic ones. However, literature has conflicting findings on the effectiveness of probabilistic classifiers in comparison to deterministic ones. In this paper, we clarify the role of randomization in building adversarially robust classifiers.Given a base hypothesis set of deterministic classifiers, we show the conditions under which a randomized ensemble outperforms the hypothesis set in adversarial risk, extending previous results.Additionally, we show that for any probabilistic binary classifier (including randomized ensembles), there exists a deterministic classifier that outperforms it. Finally, we give an explicit description of the deterministic hypothesis set that contains such a deterministic classifier for many types of commonly used probabilistic classifiers, i.e. randomized ensembles and parametric/input noise injection.

Chat is not available.