Oxford logo
[WK+19] M. Wicker and M. Kwiatkowska. Robustness of 3D Deep Learning in an Adversarial Setting. In Proceedings, Computer Vision and Pattern Recognition (CVPR) 2019. To appear. 2019. [pdf] [bib]
Downloads:  pdf pdf (1.54 MB)  bib bib
Abstract. Understanding the spatial arrangement and nature of real-world objects is of paramount importance to many complex engineering tasks, including autonomous navigation. Deep learning has revolutionized state-of-the-art performance for tasks in 3D environments; however, relatively little is known about the robustness of these approaches in an adversarial setting. The lack of comprehensive analysis makes it difficult to justify deployment of 3D deep learning models in real-world, safety-critical applications. In this work, we develop an algorithm for analysis of pointwise robustness of neural networks that operate on 3D data. We show that current approaches presented for understanding the resilience of state-of-the-art models vastly overestimate their robustness. We then use our algorithm to evaluate an array of state-of-the-art models in order to demonstrate their vulnerability to occlusion attacks. We show that, in the worst case, these networks can be reduced to 0% classification accuracy after the occlusion of at most 6.5% of the occupied input space.

QAV:

Home

People

Projects

Publications