Please use this identifier to cite or link to this item: https://dspace.iiti.ac.in/handle/123456789/4589
Title: CIIdefence: Defeating adversarial attacks by fusing class-specific image inpainting and image denoising
Authors: Gupta, Puneet
Keywords: Computer vision;Deep neural networks;Image reconstruction;Class labels;De-noising;Defence mechanisms;High confidence;Image Inpainting;Inpainting;Non-differentiable;Wavelet-based images;Image denoising
Issue Date: 2019
Publisher: Institute of Electrical and Electronics Engineers Inc.
Citation: Gupta, P., & Rahtu, E. (2019). CIIdefence: Defeating adversarial attacks by fusing class-specific image inpainting and image denoising. Paper presented at the Proceedings of the IEEE International Conference on Computer Vision, , 2019-October 6707-6716. doi:10.1109/ICCV.2019.00681
Abstract: This paper presents a novel approach for protecting deep neural networks from adversarial attacks, i.e., methods that add well-crafted imperceptible modifications to the original inputs such that they are incorrectly classified with high confidence. The proposed defence mechanism is inspired by the recent works mitigating the adversarial disturbances by the means of image reconstruction and denoising. However, unlike the previous works, we apply the reconstruction only for small and carefully selected image areas that are most influential to the current classification outcome. The selection process is guided by the class activation map responses obtained for multiple top-ranking class labels. The same regions are also the most prominent for the adversarial perturbations and hence most important to purify. The resulting inpainting task is substantially more tractable than the full image reconstruction, while still being able to prevent the adversarial attacks. Furthermore, we combine the selective image inpainting with wavelet based image denoising to produce a non differentiable layer that prevents attacker from using gradient backpropagation. Moreover, the proposed nonlinearity cannot be easily approximated with simple differentiable alternative as demonstrated in the experiments with Backward Pass Differentiable Approximation (BPDA) attack. Finally, we experimentally show that the proposed Class-specific Image Inpainting Defence (CIIDefence) is able to withstand several powerful adversarial attacks including the BPDA. The obtained results are consistently better compared to the other recent defence approaches. © 2019 IEEE.
URI: https://doi.org/10.1109/ICCV.2019.00681
https://dspace.iiti.ac.in/handle/123456789/4589
ISBN: 9781728148038
ISSN: 1550-5499
Type of Material: Conference Paper
Appears in Collections:Department of Computer Science and Engineering

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetric Badge: