Emergent Mind

Adversarial Examples Exist in Two-Layer ReLU Networks for Low Dimensional Linear Subspaces

(2303.00783)
Published Mar 1, 2023 in cs.LG , cs.CR , cs.NE , and stat.ML

Abstract

Despite a great deal of research, it is still not well-understood why trained neural networks are highly vulnerable to adversarial examples. In this work we focus on two-layer neural networks trained using data which lie on a low dimensional linear subspace. We show that standard gradient methods lead to non-robust neural networks, namely, networks which have large gradients in directions orthogonal to the data subspace, and are susceptible to small adversarial $L2$-perturbations in these directions. Moreover, we show that decreasing the initialization scale of the training algorithm, or adding $L2$ regularization, can make the trained network more robust to adversarial perturbations orthogonal to the data.

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a summary of this paper on our Pro plan:

We ran into a problem analyzing this paper.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.