Skip to content

ART Attacks

Beat Buesser edited this page Nov 29, 2020 · 43 revisions

Work in progress ...

  1. Evasion
  2. Poisoning
  3. Extraction
  4. Inference

The attack descriptions include a link to the original publication and tags describing framework-support of implementations in ART:

  • all/Numpy: implementation based on Numpy to support all frameworks
  • TensorFlow: implementation based on TensorFlow optimised for TensorFlow estimators
  • PyTorch: implementation based on PyTorch optimised for PyTorch estimators

1. Evasion Attacks

  • Auto-Attack (Croce and Hein, 2020)

    Auto-Attack runs one or more evasion attacks, defaults or provided by the user, against a classification task. Auto-Attack optimises the attack strength by only attacking correctly classified samples and by first running the untargeted version of each attack followed by running the targeted version against each possible target label.

1.1 White-box

1.2 Black-box

2. Poisoning Attacks

3. Extraction Attacks

4. Inference Attacks

4.1 Attribute Inference

4.2 Membership Inference

4.3 Model Inversion

Clone this wiki locally