Skip to yearly menu bar Skip to main content


Poster

CAMOU: Learning Physical Vehicle Camouflages to Adversarially Attack Detectors in the Wild

Yang Zhang · Hassan Foroosh · Phiip David · Boqing Gong

Great Hall BC #16

Keywords: [ synthetic simulation ] [ adversarial attack ] [ object detection ]


Abstract:

In this paper, we conduct an intriguing experimental study about the physical adversarial attack on object detectors in the wild. In particular, we learn a camouflage pattern to hide vehicles from being detected by state-of-the-art convolutional neural network based detectors. Our approach alternates between two threads. In the first, we train a neural approximation function to imitate how a simulator applies a camouflage to vehicles and how a vehicle detector performs given images of the camouflaged vehicles. In the second, we minimize the approximated detection score by searching for the optimal camouflage. Experiments show that the learned camouflage can not only hide a vehicle from the image-based detectors under many test cases but also generalizes to different environments, vehicles, and object detectors.

Live content is unavailable. Log in and register to view live content