VTFR-AT: Adversarial Training With Visual Transformation and Feature Robustness

Xiang Li, Changfei Zhao, Xinyang Deng, Wen Jiang

科研成果: 期刊稿件文章同行评审

摘要

Research on the robustness of deep neural networks to adversarial samples has grown rapidly since studies have shown that deep learning is susceptible to adversarial perturbation noise. Adversarial training is widely regarded as the most powerful defence strategy against adversarial attacks out of many defence strategies. It has been shown that the adversarial vulnerability of models is due to the learned non-robust feature in the data. However, few methods have attempted to improve adversarial training by enhancing the critical information in the data, i.e., the important region of the object. Moreover, adversarial training is prone to overfitting the model due to the overuse of training set samples. In this paper, we propose a new adversarial training framework with visual transformation and feature robustness, named VTFR-AT. The visual transformation (VT) module enhances principal information in images, weakens background information, and eliminates nuisance noise by pre-processing images. The feature robustness (FR) loss function enhances the network feature extraction partly against perturbation by constraining the feature similarity of the network on similar images. Extensive experiments have shown that the VTFR framework can substantially promote the performance of models on adversarial samples and improve the adversarial robustness and generalization capabilities. As a plug-and-play module, the proposed framework can be easily combined with various existing adversarial training methods.

源语言英语
页(从-至)3129-3140
页数12
期刊IEEE Transactions on Emerging Topics in Computational Intelligence
8
4
DOI
出版状态已出版 - 2024

指纹

探究 'VTFR-AT: Adversarial Training With Visual Transformation and Feature Robustness' 的科研主题。它们共同构成独一无二的指纹。

引用此