@article{2020arXiv200302460Y,
archivePrefix = {arXiv},
arxivId = {cs.LG/2003.02460},
author = {Yang, Yao-Yuan and Rashtchian, Cyrus and Zhang, Hongyang and Salakhutdinov, Ruslan and Chaudhuri, Kamalika},
eprint = {2003.02460},
journal = {arXiv e-prints},
keywords = {Computer Science - Cryptography and Security,Computer Science - Machine Learning,Statistics - Machine Learning},
month = {mar},
pages = {arXiv:2003.02460},
primaryClass = {cs.LG},
title = {{A Closer Look at Accuracy vs. Robustness}},
year = {2020}
}
I think this paper is worthy reading more times.
A tradeoff between robustness and accuracy may be inevitable for many classification.
Real data is $r-$separated where
Theoretically if a data distribution is $r-$separated, then there exists a robust and accurate classifier that can be obtained by rounding a locally Lipschistz function.
Figure 4 shows a pictorial example of why using a locally Lipschitz function can be just as expressive while also being robust.