Semi-Universal Adversarial Perturbations
preprintposted on 28.03.2022, 04:33 by Jordan FreconJordan Frecon, Gilles Gasso, stephane canu
The present work introduces a framework for learning and selecting semi-universal adversarial perturbations. It relies on a joint estimation of multiple universal adversarial perturbations which are chosen in an unsupervised manner depending on the sample to attack. Two algorithmic solutions, with convergence guarantees under Lipschitz continuity assumptions, are proposed to handle either small scale or large scale datasets. Numerical experiments, conducted on benchmark datasets, support its unifying aspect between universal and specific attacks as the number of perturbations grows. In addition, the learned perturbations display strong patterns indicative of the existing similarities between the training instances of different classes.
A road toward safe artificial intelligence in mobility – Raimo
Agence Nationale de la RechercheFind out more...