Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples, Anish Athalye, Nicholas Carlini, David Wagner, 2018Proceedings of the 35th International Conference on Machine Learning, Vol. 80 (PMLR) - This paper identifies gradient masking as a critical vulnerability in many proposed adversarial defenses, including input transformations, and introduces techniques like Backward Pass Differentiable Approximation (BPDA) for adaptive attacks.
Feature Squeezing: Detecting Adversarial Examples in Raw Feature Space, Weilin Xu, Youngeun Kim, Jianliang Qi, Kai-Wei Chang, David Evans, 2018International Conference on Learning Representations (ICLR) (ICLR)DOI: 10.48550/arXiv.1705.08493 - Introduces feature squeezing, an input transformation defense that uses techniques like color depth reduction and spatial smoothing to detect and potentially mitigate adversarial examples.
Synthesizing Robust Adversarial Examples, Anish Athalye, Logan Engstrom, Andrew Ilyas, Kevin Kwok, 2018Proceedings of the 35th International Conference on Machine Learning, Vol. 80 (PMLR) - Presents Expectation Over Transformation (EOT), a method for generating adversarial examples that remain effective across various random transformations, thus bypassing randomized input defenses.
Defense against Adversarial Attacks Using JPEG Compression, Gintare Karolina Dziugaite, Amna Mustafa, Cynthia Rudin, 2018NeurIPS Workshop on Challenges in Machine Learning (CIML) - Investigates the effectiveness of JPEG compression as a pre-processing input transformation defense mechanism to remove adversarial perturbations from images.