对抗攻击方法汇总(持续更新)

自从2014年Szegedy等人提出对抗样本以来,不断有研究者提出新的对抗攻击方法。本文汇总了当前已有的绝大多数算法,以抛砖引玉用,并不断更新。

Adversarial Attacks Transparency Specificity
L-BFGS White box Targeted, Non targeted
FGSM White box Targeted, Non targeted
BIM White box Targeted, Non targeted
ILCM White box Targeted
R+FGSM White box Targeted, Non targeted
AMDR White box Targeted, Non targeted
JSMA White box Targeted, Non targeted
SBA Black box Targeted, Non targeted
Hot/Cold White box Targeted
One-pixel Semi-blackbox Targeted, Non targeted
C&W White box Targeted, Non targeted
DeepFool White box Non targeted
UAP White box Non targeted
DFUAP White box Non targeted
VAE Attacks White box Targeted, Non targeted
ZOO Black box Targeted, Non targeted
UPSET Black box Targeted
ANGRI Black box Targeted
Houdini White, Black box Targeted, Non targeted
MI-FGSM White box Targeted, Non targeted
ATN White box Targeted
PGD White box Targeted
AdvGAN White box Targeted, Non targeted
Boundary Attack Black box Targeted, Non targeted
NAA Black box Non targeted
stAdv White box Targeted, Non targeted
EOT White box Targeted, Non targeted
BPDA White box Targeted, Non targeted
SPSA Black box Targeted, Non targeted
DDN White box Targeted, Non targeted
CAMOU Black box Non targeted
  • L-BFGS: Intriguing properties of neural networks
  • FGSM: Explaining and Harnessing Adversarial Examples
  • BIM & ILCM: Adversarial examples in the physical world
  • R+FGSM: Ensemble Adversarial Training: Attacks and Defenses
  • AMDR: Adversarial Manipulation of Deep Representations
  • JSMA: The Limitations of Deep Learning in Adversarial Settings
  • SBA: Practical Black-Box Attacks against Machine Learning
  • Hot/Cold: Adversarial Diversity and Hard Positive Generation
  • One-pixel: One pixel attack for fooling deep neural networks
  • C&W: Towards Evaluating the Robustness of Neural Networks
  • DeepFool: DeepFool: a simple and accurate method to fool deep neural networks
  • UAP: Universal adversarial perturbations
  • DFUAP: Generalizable Data-free Objective for Crafting Universal Adversarial Perturbations
  • VAE Attacks: Adversarial examples for generative models
  • ZOO: ZOO: Zeroth Order Optimization Based Black-box Attacks to Deep Neural Networks without Training Substitute Models
  • UPSET: UPSET and ANGRI : Breaking High Performance Image Classifiers
  • ANGRI: UPSET and ANGRI : Breaking High Performance Image Classifiers
  • Houdini: Houdini: Fooling Deep Structured Prediction Models
  • MI-FGSM: Boosting Adversarial Attacks With Momentum
  • ATN: Adversarial Transformation Networks: Learning to Generate Adversarial Examples
  • PGD: Towards Deep Learning Models Resistant to Adversarial Attacks
  • AdvGAN: ## Generating Adversarial Examples with Adversarial Networks
  • Boundary Attack: Decision-Based Adversarial Attacks: Reliable Attacks Against Black-Box Machine Learning Models
  • NAA: ## Generating Natural Adversarial Examples
  • stAdv: Spatially Transformed Adversarial Examples
  • EOT: Synthesizing Robust Adversarial Examples
  • BPDA: Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples
  • SPSA: Multivariate stochastic approximation using a simultaneous perturbation gradient approximation
  • DDN: Decoupling Direction and Norm for Efficient Gradient-Based L2 Adversarial Attacks and Defenses
  • CAMOU: CAMOU: Learning Physical Vehicle Camouflages to Adversarially Attack Detectors in the Wild

参考

[1] Akhtar N, Mian A. Threat of adversarial attacks on deep learning in computer vision: A survey[J]. IEEE Access, 2018, 6: 14410-14430.
[2] Yuan X, He P, Zhu Q, et al. Adversarial examples: Attacks and defenses for deep learning[J]. IEEE transactions on neural networks and learning systems, 2019, 30(9): 2805-2824.
[3] Wiyatno R R, Xu A, Dia O, et al. Adversarial Examples in Modern Machine Learning: A Review[J]. arXiv preprint arXiv:1911.05268, 2019.

你可能感兴趣的:(对抗攻击方法汇总(持续更新))