TY - JOUR
T1 - Understanding adversarial training
T2 - Increasing local stability of supervised models through robust optimization
AU - Shaham, Uri
AU - Yamada, Yutaro
AU - Negahban, Sahand
N1 - Publisher Copyright:
© 2018 Elsevier B.V.
PY - 2018/9/13
Y1 - 2018/9/13
N2 - We show that adversarial training of supervised learning models is in fact a robust optimization procedure. To do this, we establish a general framework for increasing local stability of supervised learning models using robust optimization. The framework is general and broadly applicable to differentiable non-parametric models, e.g., Artificial Neural Networks (ANNs). Using an alternating minimization-maximization procedure, the loss of the model is minimized with respect to perturbed examples that are generated at each parameter update, rather than with respect to the original training data. Our proposed framework generalizes adversarial training, as well as previous approaches for increasing local stability of ANNs. Experimental results reveal that our approach increases the robustness of the network to existing adversarial examples, while making it harder to generate new ones. Furthermore, our algorithm improves the accuracy of the networks also on the original test data.
AB - We show that adversarial training of supervised learning models is in fact a robust optimization procedure. To do this, we establish a general framework for increasing local stability of supervised learning models using robust optimization. The framework is general and broadly applicable to differentiable non-parametric models, e.g., Artificial Neural Networks (ANNs). Using an alternating minimization-maximization procedure, the loss of the model is minimized with respect to perturbed examples that are generated at each parameter update, rather than with respect to the original training data. Our proposed framework generalizes adversarial training, as well as previous approaches for increasing local stability of ANNs. Experimental results reveal that our approach increases the robustness of the network to existing adversarial examples, while making it harder to generate new ones. Furthermore, our algorithm improves the accuracy of the networks also on the original test data.
KW - Adversarial examples
KW - Deep learning
KW - Non-parametric supervised models
KW - Robust optimization
UR - http://www.scopus.com/inward/record.url?scp=85047515976&partnerID=8YFLogxK
U2 - 10.1016/j.neucom.2018.04.027
DO - 10.1016/j.neucom.2018.04.027
M3 - ???researchoutput.researchoutputtypes.contributiontojournal.article???
AN - SCOPUS:85047515976
SN - 0925-2312
VL - 307
SP - 195
EP - 204
JO - Neurocomputing
JF - Neurocomputing
ER -