AutoAttack
AutoAttack for Adversarial Robustness
Introduction
Adversarial training is about robustify a neural network against adversarial attacks.
- More details: here.
- Link to AutoAttack.
Key insights
Authors do not argue that AutoAttack [1] is the ultimate adversarial attack but rather that it should become the minimal test for any new defense, since it reliably reaches good performance in all tested models, without any hyperparameter tuning and at a relatively low computational cost.
3 weaknesses of PGD:
- Fixed step size: suboptimal, even for convex problems this does not guarantee convergence, and the performance of the algorithm is highly influenced by the choice of the value. [2]
- Agnostic of the budget: The loss plateaus after a few iterations, except for extremely small step sizes, which however do not translate into better results. Judging the strength of an attack by the number of iterations is misleading. [3]
- Unaware of the trend: Does not consider whether the optimization is evolving successfully and is not able to react of this. Authors present an automatic scheme fixing this issue.
References
- [1] [2003.01690] Reliable Evaluation of Adversarial Robustness with an Ensemble of Diverse Parameter-free Attacks
- [2] [1810.12042] Logit Pairing Methods Can Fool Gradient-Based Attacks
- [3] [1902.06705] On Evaluating Adversarial Robustness
Enjoy Reading This Article?
Here are some more articles you might like to read next: