STDatav2: Accessing Efficient Black-Box Stealing for Adversarial Attacks

Xuxiang Sun, Gong Cheng, Hongda Li, Chunbo Lang, Junwei Han

Research output: Contribution to journalArticlepeer-review

Abstract

On account of the extreme settings, stealing the black-box model without its training data is difficult in practice. On this topic, along the lines of data diversity, this paper substantially makes the following improvements based on our conference version (dubbed STDatav1, short for Surrogate Training Data). First, to mitigate the undesirable impacts of the potential mode collapse while training the generator, we propose the joint-data optimization scheme, which utilizes both the synthesized data and the proxy data to optimize the surrogate model. Second, we propose the self-conditional data synthesis framework, an interesting effort that builds the pseudo-class mapping framework via grouping class information extraction to hold the class-specific constraints while holding the diversity. Within this new framework, we inherit and integrate the class-specific constraints of STDatav1 and design a dual cross-entropy loss to fit this new framework. Finally, to facilitate comprehensive evaluations, we perform experiments on four commonly adopted datasets, and a total of eight kinds of models are employed. These assessments witness the considerable performance gains compared to our early work and demonstrate the competitive ability and promising potential of our approach.

Original languageEnglish
Pages (from-to)2429-2445
Number of pages17
JournalIEEE Transactions on Pattern Analysis and Machine Intelligence
Volume47
Issue number4
DOIs
StatePublished - 2025

Keywords

  • Black-box attacks
  • joint-data optimization
  • model stealing
  • self-conditional data synthesis
  • surrogate training data (STData)

Fingerprint

Dive into the research topics of 'STDatav2: Accessing Efficient Black-Box Stealing for Adversarial Attacks'. Together they form a unique fingerprint.

Cite this