论文标题

对抗性鲁棒性:从自我监督的预训练到微调

Adversarial Robustness: From Self-Supervised Pre-Training to Fine-Tuning

论文作者

Chen, Tianlong, Liu, Sijia, Chang, Shiyu, Cheng, Yu, Amini, Lisa, Wang, Zhangyang

论文摘要

从自我划分的验证模型普遍用于更快或更准确的下游任务进行微调。但是,从训练中获得鲁棒性并未探索。我们将对抗性训练引入自学训练,以首次提供通用的强大预训练模型。我们发现这些健壮的预训练模型可以通过两种方式使后续的微调受益:i)提高最终模型鲁棒性; ii)保留计算成本,如果进行对抗微调。我们进行了广泛的实验,以证明所提出的框架达到了较大的性能边距(例如,与常规的端到端对抗性训练基线相比,在CIFAR-10数据集上,稳健精度为3.83%,标准精度为1.3%)。此外,我们发现不同的自我监管的预训练模型具有多种对抗性脆弱性。它启发了我们整合几项训练预处理的任务,这会更多地提高鲁棒性。我们的整体策略在鲁棒精度上进一步提高了3.59%,同时保持了CIFAR-10的标准精度略高。我们的代码可从https://github.com/tamu-vita/adv-ss-pretraining获得。

Pretrained models from self-supervision are prevalently used in fine-tuning downstream tasks faster or for better accuracy. However, gaining robustness from pretraining is left unexplored. We introduce adversarial training into self-supervision, to provide general-purpose robust pre-trained models for the first time. We find these robust pre-trained models can benefit the subsequent fine-tuning in two ways: i) boosting final model robustness; ii) saving the computation cost, if proceeding towards adversarial fine-tuning. We conduct extensive experiments to demonstrate that the proposed framework achieves large performance margins (eg, 3.83% on robust accuracy and 1.3% on standard accuracy, on the CIFAR-10 dataset), compared with the conventional end-to-end adversarial training baseline. Moreover, we find that different self-supervised pre-trained models have a diverse adversarial vulnerability. It inspires us to ensemble several pretraining tasks, which boosts robustness more. Our ensemble strategy contributes to a further improvement of 3.59% on robust accuracy, while maintaining a slightly higher standard accuracy on CIFAR-10. Our codes are available at https://github.com/TAMU-VITA/Adv-SS-Pretraining.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源