Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Protecting Intellectual Property of Generative Adversarial Networks from Ambiguity Attack

About

Ever since Machine Learning as a Service (MLaaS) emerges as a viable business that utilizes deep learning models to generate lucrative revenue, Intellectual Property Right (IPR) has become a major concern because these deep learning models can easily be replicated, shared, and re-distributed by any unauthorized third parties. To the best of our knowledge, one of the prominent deep learning models - Generative Adversarial Networks (GANs) which has been widely used to create photorealistic image are totally unprotected despite the existence of pioneering IPR protection methodology for Convolutional Neural Networks (CNNs). This paper therefore presents a complete protection framework in both black-box and white-box settings to enforce IPR protection on GANs. Empirically, we show that the proposed method does not compromise the original GANs performance (i.e. image generation, image super-resolution, style transfer), and at the same time, it is able to withstand both removal and ambiguity attacks against embedded watermarks.

Ding Sheng Ong, Chee Seng Chan, Kam Woh Ng, Lixin Fan, Qiang Yang• 2021

Related benchmarks

TaskDatasetResultRank
Watermark RestorationCIFAR-100 original (test)
Clean Similarity100
13
Watermark similarity restorationResNet-18 watermarked
Clean Similarity1
9
Watermark similarity restorationEfficientNet watermarked
Clean Similarity100
9
Watermark similarity restorationInception V3
Clean Similarity Score1
9
Watermark similarity restorationDenseNet
Clean Similarity1
9
Showing 5 of 5 rows

Other info

Follow for update