A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2020; you can also visit the original URL.
The file type is application/pdf
.
Detecting and Diagnosing Adversarial Images with Class-Conditional Capsule Reconstructions
[article]
2020
arXiv
pre-print
Adversarial examples raise questions about whether neural network models are sensitive to the same visual features as humans. In this paper, we first detect adversarial examples or otherwise corrupted images based on a class-conditional reconstruction of the input. To specifically attack our detection mechanism, we propose the Reconstructive Attack which seeks both to cause a misclassification and a low reconstruction error. This reconstructive attack produces undetected adversarial examples
arXiv:1907.02957v2
fatcat:xcxwzrth5jhmnbtrxcxvfuy2gu