TY - JOUR
T1 - Why adversarial reprogramming works, when it fails, and how to tell the difference
AU - Zheng, Yang
AU - Feng, Xiaoyi
AU - Xia, Zhaoqiang
AU - Jiang, Xiaoyue
AU - Demontis, Ambra
AU - Pintor, Maura
AU - Biggio, Battista
AU - Roli, Fabio
N1 - Publisher Copyright:
© 2023 Elsevier Inc.
PY - 2023/6
Y1 - 2023/6
N2 - Adversarial reprogramming allows repurposing a machine-learning model to perform a different task. For example, a model trained to recognize animals can be reprogrammed to recognize digits by embedding an adversarial program in the digit images provided as input. Recent work has shown that adversarial reprogramming may not only be used to abuse machine-learning models provided as a service, but also beneficially, to improve transfer learning when training data is scarce. However, the factors affecting its success are still largely unexplained. In this work, we develop a first-order linear model of adversarial reprogramming to show that its success inherently depends on the size of the average input gradient, which grows when input gradients are more aligned, and when inputs have higher dimensionality. The results of our experimental analysis, involving fourteen distinct reprogramming tasks, show that the above factors are correlated with the success and the failure of adversarial reprogramming.
AB - Adversarial reprogramming allows repurposing a machine-learning model to perform a different task. For example, a model trained to recognize animals can be reprogrammed to recognize digits by embedding an adversarial program in the digit images provided as input. Recent work has shown that adversarial reprogramming may not only be used to abuse machine-learning models provided as a service, but also beneficially, to improve transfer learning when training data is scarce. However, the factors affecting its success are still largely unexplained. In this work, we develop a first-order linear model of adversarial reprogramming to show that its success inherently depends on the size of the average input gradient, which grows when input gradients are more aligned, and when inputs have higher dimensionality. The results of our experimental analysis, involving fourteen distinct reprogramming tasks, show that the above factors are correlated with the success and the failure of adversarial reprogramming.
KW - Adversarial machine learning
KW - Adversarial reprogramming
KW - Neural networks
KW - Transfer learning
UR - http://www.scopus.com/inward/record.url?scp=85149745021&partnerID=8YFLogxK
U2 - 10.1016/j.ins.2023.02.086
DO - 10.1016/j.ins.2023.02.086
M3 - 文章
AN - SCOPUS:85149745021
SN - 0020-0255
VL - 632
SP - 130
EP - 143
JO - Information Sciences
JF - Information Sciences
ER -