Towards Deep Learning Models Resistant to Adversarial Attacks, Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, Adrian Vladu, 2018International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.1706.06083 - Introduces PGD adversarial training, a foundational and widely adopted defense technique for improving model robustness.
Certified Adversarial Robustness via Randomized Smoothing, Jeremy Cohen, Elan Rosenfeld, Zico Kolter, 2019Proceedings of the 36th International Conference on Machine Learning (ICML), Vol. 97 (PMLR)DOI: 10.48550/arXiv.1902.02918 - Presents randomized smoothing, a method to achieve provable robustness certificates for deep learning models against adversarial attacks.