Samples on Thin Ice: Re-evaluating Adversarial Pruning of Neural Networks
Giorgio Piras
First
;Maura PintorSecond
;Ambra DemontisPenultimate
;Battista BiggioLast
2023-01-01
Abstract
Neural network pruning has shown to be an effective technique for reducing the network size, trading desirable properties like generalization and robustness to adversarial attacks for higher sparsity. Recent work has claimed that adversarial pruning methods can produce sparse networks while also preserving robustness to adversarial examples. In this work, we first re-evaluate three state-of-the-art adversarial pruning methods, showing that their robustness was indeed overestimated. We then compare pruned and dense versions of the same models, discovering that samples on thin ice, i.e., closer to the unpruned model’s decision boundary, are typically misclassified after pruning. We conclude by discussing how this intuition may lead to designing more effective adversarial pruning methods in future work.File | Size | Format | |
---|---|---|---|
2310.08073.pdf open access
Type: versione pre-print
Size 380.56 kB
Format Adobe PDF
|
380.56 kB | Adobe PDF | View/Open |
Samples_on_Thin_Ice_Re-Evaluating_Adversarial_Pruning_of_Neural_Networks.pdf Solo gestori archivio
Type: versione editoriale
Size 480.94 kB
Format Adobe PDF
|
480.94 kB | Adobe PDF | & nbsp; View / Open Request a copy |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.