papers AI Learner
The Github is limit! Click to go to the new site.

Robustness of 3D Deep Learning in an Adversarial Setting

2019-04-01
Matthew Wicker, Marta Kwiatkowska

Abstract

Understanding the spatial arrangement and nature of real-world objects is of paramount importance to many complex engineering tasks, including autonomous navigation. Deep learning has revolutionized state-of-the-art performance for tasks in 3D environments; however, relatively little is known about the robustness of these approaches in an adversarial setting. The lack of comprehensive analysis makes it difficult to justify deployment of 3D deep learning models in real-world, safety-critical applications. In this work, we develop an algorithm for analysis of pointwise robustness of neural networks that operate on 3D data. We show that current approaches presented for understanding the resilience of state-of-the-art models vastly overestimate their robustness. We then use our algorithm to evaluate an array of state-of-the-art models in order to demonstrate their vulnerability to occlusion attacks. We show that, in the worst case, these networks can be reduced to 0% classification accuracy after the occlusion of at most 6.5% of the occupied input space.

Abstract (translated by Google)
URL

http://arxiv.org/abs/1904.00923

PDF

http://arxiv.org/pdf/1904.00923


Similar Posts

Comments