Diffusion Models-Based Purification for Common Corruptions on Robust 3D Object Detection
Author:
Cai Mumuxin1, Wang Xupeng2ORCID, Sohel Ferdous3ORCID, Lei Hang1
Affiliation:
1. School of Information and Software Engineering, The University of Electronic Science and Technology of China, Chengdu 610054, China 2. Laboratory Of Intelligent Collaborative Computing, The University of Electronic Science and Technology of
China, Chengdu 610054, China 3. School of Information Technology, Murdoch University, Perth, WA 6150, Australia
Abstract
LiDAR sensors have been shown to generate data with various common corruptions, which seriously affect their applications in 3D vision tasks, particularly object detection. At the same time, it has been demonstrated that traditional defense strategies, including adversarial training, are prone to suffering from gradient confusion during training. Moreover, they can only improve their robustness against specific types of data corruption. In this work, we propose LiDARPure, which leverages the powerful generation ability of diffusion models to purify corruption in the LiDAR scene data. By dividing the entire scene into voxels to facilitate the processes of diffusion and reverse diffusion, LiDARPure overcomes challenges induced from adversarial training, such as sparse point clouds in large-scale LiDAR data and gradient confusion. In addition, we utilize the latent geometric features of a scene as a condition to assist the generation of diffusion models. Detailed experiments show that LiDARPure can effectively purify 19 common types of LiDAR data corruption. Further evaluation results demonstrate that it can improve the average precision of 3D object detectors to an extent of 20% in the face of data corruption, much higher than existing defence strategies.
Funder
National Natural Science Foundation of China Sichuan Provincial Research Plan Project
Reference50 articles.
1. Point-cloud based 3D object detection and classification methods for self-driving applications: A survey and taxonomy;Fernandes;Inf. Fusion,2021 2. Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I.J., and Fergus, R. (2014, January 14–16). Intriguing properties of neural networks. Proceedings of the International Conference on Learning Representations, Banff, AB, Canada. 3. Moosavi-Dezfooli, S.M., Fawzi, A., and Frossard, P. (2016, January 27–30). DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA. 4. Tu, J., Ren, M., Manivasagam, S., Liang, M., Yang, B., Du, R., Cheng, F., and Urtasun, R. (2020, January 13–19). Physically Realizable Adversarial Examples for LiDAR Object Detection. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA. 5. Adversarial point cloud perturbations against 3D object detection in autonomous driving systems;Wang;Neurocomputing,2021
|
|