Efficient proximal mapping of the 1-path-norm of shallow networks

Fabian Latorre*, Paul Rolland, Nadav Hallak, Volkan Cevher

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

We demonstrate two new important properties of the 1-path-norm of shallow neural networks. First, despite its non-smoothness and non-convexity it allows a closed form proximal operator which can be efficiently computed, allowing the use of stochastic proximal-gradient-type methods for regularized empirical risk minimization. Second, when the activation functions is differentiable, it provides an upper bound on the Lipschitz con_stant of the network. Such bound is tighter than the trivial layer-wise product of Lipschitz con_stants, motivating its use for training networks robust to adversarial perturbations. In practical experiments we illustrate the advantages of us_ing the proximal mapping and we compare the robustness-accuracy trade-off induced by the 1- path-norm, L1-norm and layer-wise constraints on the Lipschitz constant (Parseval networks).

Original languageEnglish
Title of host publication37th International Conference on Machine Learning, ICML 2020
EditorsHal Daume, Aarti Singh
PublisherInternational Machine Learning Society (IMLS)
Pages5607-5617
Number of pages11
ISBN (Electronic)9781713821120
StatePublished - 2020
Externally publishedYes
Event37th International Conference on Machine Learning, ICML 2020 - Virtual, Online
Duration: 13 Jul 202018 Jul 2020

Publication series

Name37th International Conference on Machine Learning, ICML 2020
VolumePartF168147-8

Conference

Conference37th International Conference on Machine Learning, ICML 2020
CityVirtual, Online
Period13/07/2018/07/20

Fingerprint

Dive into the research topics of 'Efficient proximal mapping of the 1-path-norm of shallow networks'. Together they form a unique fingerprint.

Cite this