Abstract
Machine learning (ML) models are applied in an increasing variety of domains. The availability of large amounts of data and computational resources encourages the development of ever more complex and valuable models. These models are considered the intellectual property of the legitimate parties who have trained them, which makes their protection against stealing, illegitimate redistribution, and unauthorized application an urgent need. Digital watermarking presents a strong mechanism for marking model ownership and, thereby, offers protection against those threats. This work presents a taxonomy identifying and analyzing different classes of watermarking schemes for ML models. It introduces a unified threat model to allow structured reasoning on and comparison of the effectiveness of watermarking methods in different scenarios. Furthermore, it systematizes desired security requirements and attacks against ML model watermarking. Based on that framework, representative literature from the field is surveyed to illustrate the taxonomy. Finally, shortcomings and general limitations of existing approaches are discussed, and an outlook on future research directions is given.
Subject
Artificial Intelligence,Information Systems,Computer Science (miscellaneous)
Reference61 articles.
1. Turning Your Weakness into a Strength: Watermarking Deep Neural Networks by Backdooring;Adi,2018
2. Neural Network Laundering: Removing Black-Box Backdoor Watermarks from Deep Neural Networks;Aiken;Comput. Security,2021
3. Hacking Smart Machines with Smarter Ones: How to Extract Meaningful Data from Machine Learning Classifiers;Ateniese,2013
4. Pruning Algorithms of Neural Networks—A Comparative Study;Augasta;Open Comp. Sci.,2013
5. Csi Neural Network: Using Side-Channels to Recover Your Artificial Neural Network Information;Batina,2018
Cited by
23 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献