BlindSpot: Watermarking through fairness

Lounici, Sofiane; Önen, Melek; Ermis, Orhan; Trabelsi, Slim
IH&MMSEC 2022, 10th ACM Workshop on Information Hiding and Multimedia Security, June 27-28, 2022, Santa Barbara, California, USA

With the increasing development of machine learning models in daily businesses, a strong need for intellectual property protection arised. For this purpose, current works suggest to leverage backdoor techniques to embed a watermark into the model, by overfitting to a set of particularly crafted and secret input-output pairs called triggers. By sending verification queries containing triggers, the model owner can analyse the behavior of any suspect model on the queries to claim its ownership. However, when it comes to scenarios where frequent monitoring is needed, the computational overhead of these verification queries in terms of volume demonstrates that backdoor-based watermarking appears to be too sensitive to outlier detection attacks and cannot guarantee the secrecy of the triggers.

To solve this issue, we introduce BlindSpot, to watermark machine learning models through fairness. Our trigger-less approach is compatible with a high number of verification queries while being robust to outlier detection attacks. We show on Fashion-MNIST and CIFAR-10 datasets that BlindSpot is efficiently watermarking models while robust to outlier detection attacks, at a performance cost on the accuracy of 2%.


DOI
Type:
Conférence
City:
Santa Barbara
Date:
2022-06-27
Department:
Sécurité numérique
Eurecom Ref:
6890
Copyright:
© ACM, 2022. This is the author's version of the work. It is posted here by permission of ACM for your personal use. Not for redistribution. The definitive version was published in IH&MMSEC 2022, 10th ACM Workshop on Information Hiding and Multimedia Security, June 27-28, 2022, Santa Barbara, California, USA https://doi.org/10.1145/3531536.3532950

PERMALINK : https://www.eurecom.fr/publication/6890