Back to Search Start Over

Copycat CNN: Are Random Non-Labeled Data Enough to Steal Knowledge from Black-box Models?

Authors :
Correia-Silva, Jacson Rodrigues
Berriel, Rodrigo F.
Badue, Claudine
De Souza, Alberto F.
Oliveira-Santos, Thiago
Source :
Pattern Recognition 113 (2021) 107830
Publication Year :
2021

Abstract

Convolutional neural networks have been successful lately enabling companies to develop neural-based products, which demand an expensive process, involving data acquisition and annotation; and model generation, usually requiring experts. With all these costs, companies are concerned about the security of their models against copies and deliver them as black-boxes accessed by APIs. Nonetheless, we argue that even black-box models still have some vulnerabilities. In a preliminary work, we presented a simple, yet powerful, method to copy black-box models by querying them with natural random images. In this work, we consolidate and extend the copycat method: (i) some constraints are waived; (ii) an extensive evaluation with several problems is performed; (iii) models are copied between different architectures; and, (iv) a deeper analysis is performed by looking at the copycat behavior. Results show that natural random images are effective to generate copycats for several problems.<br />Comment: The code is available at https://github.com/jeiks/Stealing_DL_Models

Details

Database :
arXiv
Journal :
Pattern Recognition 113 (2021) 107830
Publication Type :
Report
Accession number :
edsarx.2101.08717
Document Type :
Working Paper
Full Text :
https://doi.org/10.1016/j.patcog.2021.107830