Deep machine learning models, e.g., image classifier, are increasingly deployed in the wild to provide services to users. Adversaries are shown capable of stealing the knowledge of these models by sending inference queries and then training substitute models based on query results. The availability and quality of adversarial query inputs are undoubtedly crucial in the stealing process. The recent prior art demonstrates the feasibility of replacing real data by exploring the synthetic adversarial queries, so called data-free attacks, under strong adversarial assumptions, i.e., the deployed classier returns not only class labels but also class probabilities. In this paper, we consider a general adversarial model and propose an effective data-free stealing algorithm, TandemGAN, which not only explores synthetic queries but also explicitly exploits the high quality ones. The core of TandemGAN is composed of (i) substitute model which imitates the target model through synthetic queries and their inferred labels; and (ii) a tandem generator consisting of two networks, {\$}{\$}{\backslash}mathcal {\{}G{\}}{\_}x{\$}{\$}and {\$}{\$}{\backslash}mathcal {\{}G{\}}{\_}e{\$}{\$}, which first explores the synthetic data space via {\$}{\$}{\backslash}mathcal {\{}G{\}}{\_}x{\$}{\$}and then exploits high-quality examples via {\$}{\$}{\backslash}mathcal {\{}G{\}}{\_}e{\$}{\$}to maximize the knowledge transfer from the target to the substitute model. Our results on four datasets show that the accuracy of our trained substitute model ranges between 96--67{\%} of the target model and outperforms the existing state-of-the-art data-free model stealing approach by up to 2.5X.

Exploring and Exploiting Data-Free Model Stealing

Robert Birke;
2023-01-01

Abstract

Deep machine learning models, e.g., image classifier, are increasingly deployed in the wild to provide services to users. Adversaries are shown capable of stealing the knowledge of these models by sending inference queries and then training substitute models based on query results. The availability and quality of adversarial query inputs are undoubtedly crucial in the stealing process. The recent prior art demonstrates the feasibility of replacing real data by exploring the synthetic adversarial queries, so called data-free attacks, under strong adversarial assumptions, i.e., the deployed classier returns not only class labels but also class probabilities. In this paper, we consider a general adversarial model and propose an effective data-free stealing algorithm, TandemGAN, which not only explores synthetic queries but also explicitly exploits the high quality ones. The core of TandemGAN is composed of (i) substitute model which imitates the target model through synthetic queries and their inferred labels; and (ii) a tandem generator consisting of two networks, {\$}{\$}{\backslash}mathcal {\{}G{\}}{\_}x{\$}{\$}and {\$}{\$}{\backslash}mathcal {\{}G{\}}{\_}e{\$}{\$}, which first explores the synthetic data space via {\$}{\$}{\backslash}mathcal {\{}G{\}}{\_}x{\$}{\$}and then exploits high-quality examples via {\$}{\$}{\backslash}mathcal {\{}G{\}}{\_}e{\$}{\$}to maximize the knowledge transfer from the target to the substitute model. Our results on four datasets show that the accuracy of our trained substitute model ranges between 96--67{\%} of the target model and outperforms the existing state-of-the-art data-free model stealing approach by up to 2.5X.
2023
European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Database (PKDD and ECML combined from 2008)
Turin, Italy
18-22 Sep
European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Database (PKDD and ECML combined from 2008)
Springer
20
35
978-3-031-43423-5
Chi Hong, Jiyue Huang, Robert Birke, Lydia Y. Chen
File in questo prodotto:
File Dimensione Formato  
Data-free Model Stealing.pdf

Accesso aperto

Descrizione: preprint
Tipo di file: PREPRINT (PRIMA BOZZA)
Dimensione 877.19 kB
Formato Adobe PDF
877.19 kB Adobe PDF Visualizza/Apri
978-3-031-43424-2_2.pdf

Accesso riservato

Tipo di file: PDF EDITORIALE
Dimensione 1.03 MB
Formato Adobe PDF
1.03 MB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2318/1923515
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 3
  • ???jsp.display-item.citation.isi??? 1
social impact