Deep Learning systems, able to achieve significant breakthroughs in many fields, including computer vision and speech recognition, are not inherently secure. Adversarial attacks on computer vision models can craft slightly perturbed inputs that exploit the models’ multi-dimensional boundary shape to dramatically reduce their performance without compromising the perception that human beings have of such input. In this work, we present Trust-No-Pixel, a novel plug-and-play strategy to harden neural network image classifiers from adversarial attacks, based on a massive inpainting strategy. The inpainting technique of our defense performs a total erase of the input image and its reconstruction from scratch. Our experiments show Trust-No- Pixel improved accuracy against the more challenging type of such attacks, namely the white box adversarial attacks. Moreover, an exhaustive comparison of our technique against state-of- the-art approaches taken from academic literature confirmed the solid defense performances of Trust-No-Pixel under a wide variety of scenarios, including different attacks and attacked network architectures.

Trust-No-Pixel: A Remarkably Simple Defense against Adversarial Attacks Based on Massive Inpainting

Stefano Samele;Matteo Matteucci
2022-01-01

Abstract

Deep Learning systems, able to achieve significant breakthroughs in many fields, including computer vision and speech recognition, are not inherently secure. Adversarial attacks on computer vision models can craft slightly perturbed inputs that exploit the models’ multi-dimensional boundary shape to dramatically reduce their performance without compromising the perception that human beings have of such input. In this work, we present Trust-No-Pixel, a novel plug-and-play strategy to harden neural network image classifiers from adversarial attacks, based on a massive inpainting strategy. The inpainting technique of our defense performs a total erase of the input image and its reconstruction from scratch. Our experiments show Trust-No- Pixel improved accuracy against the more challenging type of such attacks, namely the white box adversarial attacks. Moreover, an exhaustive comparison of our technique against state-of- the-art approaches taken from academic literature confirmed the solid defense performances of Trust-No-Pixel under a wide variety of scenarios, including different attacks and attacked network architectures.
2022
Proceedings of the 2022 International Joint Conference on Neural Networks (IJCNN)
Adversarial attacks, Inpainting, DeepFill
File in questo prodotto:
File Dimensione Formato  
2022087418.pdf

Accesso riservato

: Publisher’s version
Dimensione 4.63 MB
Formato Adobe PDF
4.63 MB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11311/1220501
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? 0
social impact