Perturbing Across the Feature Hierarchy to Improve Standard and Strict Blackbox Attack Transferability

Nathan Inkawhich, Kevin J Liang, Binghui Wang, Matthew Inkawhich, Lawrence Carin, Yiran Chen

Research output: Contribution to journalArticlepeer-review

Abstract

We consider the blackbox transfer-based targeted adversarial attack threat model in the realm of deep neural network (DNN) image classifiers. Rather than focusing on crossing decision boundaries at the output layer of the source model, our method perturbs representations throughout the extracted feature hierarchy to resemble other classes. We design a flexible attack framework that allows for multi-layer perturbations and demonstrates state-of-the-art targeted transfer performance between ImageNet DNNs. We also show the superiority of our feature space methods under a relaxation of the common assumption that the source and target models are trained on the same dataset and label space, in some instances achieving a $10\times$ increase in targeted success rate relative to other blackbox transfer methods. Finally, we analyze why the proposed methods outperform existing attack strategies and show an extension of the method in the case when limited queries to the blackbox model are allowed.
Original languageEnglish (US)
JournalArxiv preprint
StatePublished - Apr 29 2020
Externally publishedYes

Keywords

  • cs.CR
  • cs.LG
  • stat.ML

Fingerprint

Dive into the research topics of 'Perturbing Across the Feature Hierarchy to Improve Standard and Strict Blackbox Attack Transferability'. Together they form a unique fingerprint.

Cite this