|

Which model architecture is the best in adversarial defense?

Introduction

The upcoming topic in the AI security @ CVPR ’23 series is adversarial defense: scientific work that doesn’t present new attacks, but rather focuses on bolstering the defenses against existing ones. To kickstart the shift from attacks to defense in the series, this week I thought I’d create a small tool to compare adversarial defense of various computer vision model architectures based on an interesting CVPR ’23 paper:

A. Liu et al.: Exploring the Relationship Between Architectural Design and Adversarially Robust Generalization (CVPR ’23)

A. Liu et al. have evaluated the defenses of 20 computer vision model architectures against adversarial attacks based on various \(l\)-norms. Recall that the \(l\)-norm used in the attack shapes how the attack looks. \(l_2\)– and \(l_{\infty}\)-norm attacks are the classic perturbations across the entire image. By contrast, sparse attacks arbitrarily modifying only a few select pixels employ \(l_0\)-norm restrictions. The paper brings many interesting insights into the robustness to adversarial attacks. If you’re interested in this topic, I recommend you read the paper. You might also want to check out http://robust.art, a robustness benchmark created by the authors.

Comparing architectural robustness

In my opinion, the real gem is Table 1, which features accuracy results of 20 model architectures under various attacks on CIFAR-10. The table reports vanilla accuracy (standard accuracy in the classic, non-adversarial setting), clean accuracy (accuracy on clean images in the adversarial setting), robust accuracy on images perturbed by adversarial attacks, and worst case accuracy, the lower bound on robustness under multiple adversarial attacks. The authors evaluate two types of attacks, AutoAttack (AA) and projected gradient descent (PGD) attacks. The attacks on CIFAR-10 employ \(l_1\) (\(\epsilon = 40.0\)), \(l_2\) (\(\epsilon = 8.0\)), and \(l_{\infty}\) (\(\epsilon = \frac{8}{255}\)) norms.

So, without further ado: the small tool I advertised in the first paragraph is a sortable version of accuracy under adversarial attacks collected by A. Liu et al. You can sort the table by multiple columns by holding Shift and then click on the column sorting arrows in the order you want. I hope this provides useful comparisons and enables even more insight into the adversarial defense of various computer vision model architectures.

ArchParams
(M)
VanillaClean PGD-\(l_{\infty}\)AA-\(l_{\infty}\)PGD-\(l_2\)PGD-\(l_1\)Worst case
PvTv212.4088.3475.9946.4838.1835.7746.1433.54
CoAtNet16.9990.7377.7348.2739.8533.8042.3032.17
ViT9.7886.7378.7646.0238.0030.8639.2729.24
CPVT9.4990.3478.5745.0236.7330.1539.2228.47
ViTAE23.1888.2475.4240.5333.2229.6740.0228.13
MLP-Mixer0.6883.4362.8638.9331.8129.2736.5027.42
PoolFormer11.3989.2673.6646.3338.9328.8434.3227.36
CCT3.7692.2781.2349.2140.9728.2934.5926.82
VGG14.7294.0184.3050.8741.6626.7831.4825.32
Swin Trans.27.4291.5880.4448.6141.3126.5830.4725.04
LeViT6.6789.0177.1047.1639.8726.2829.5825.04
MobileViT5.0091.4777.5249.5141.5026.9629.3524.41
BoTNet18.8294.1680.7651.2942.9525.8427.3823.15
WideResNet55.8596.4789.5455.1744.1322.5523.6820.88
DenseNet1.1294.4283.2353.0644.0222.5521.8719.48
PreActResNet23.5095.8687.9654.8545.8118.6016.4615.11
CeiT5.5685.2471.5536.2028.0215.3116.7714.35
ResNet23.5295.6087.9254.1845.4017.5215.9014.32
ResNeXt9.1295.6487.1251.5142.6615.0713.6412.18
CvT19.5487.8173.7641.3633.6712.759.258.76
Table 1: A sortable table of accuracy results for 20 computer vision model architectures on CIFAR-10. Data collected from A. Liu et al.: Exploring the Relationship Between Architectural Design and Adversarially Robust Generalization (CVPR ’23).

Subscribe

Enjoying the blog? Subscribe to receive blog updates, post notifications, and monthly post summaries by e-mail.

Similar Posts