Federated learning (FL) has become an emerging
distributed framework to build deep learning models with
collaborative efforts from multiple participants. Consequently,
copyright protection of FL deep model is urgently required
because too many participants have access to the joint-trained
model. Recently, encrypted FL framework is developed to
address data leakage issue when central node is not fully
trustable. This encryption process has made existing DL model
watermarking schemes impossible to embed watermark at
the central node. In this paper, we propose a novel clientside federated learning watermarking method to tackle the
model verification issue under the encrypted FL framework. In
specific, we design a backdoor-based watermarking scheme to
allow model owners to embed their pre-designed noise patterns
into the FL deep model. Thus, our method provides reliable
copyright protection while ensuring the data privacy because
the central node has no access to the encrypted gradient
information. The experimental results have demonstrated the
efficiency of our method in terms of both FL model performance
and watermarking robustness.
Funding
National Natural Science Foundation of China (61602527, 61772555, 61772553, U1734208)
Natural Science Foundation of Hunan Province, China (2020JJ4746)
History
School
Science
Department
Computer Science
Published in
2021 IEEE International Conference on Systems, Man, and Cybernetics (SMC)
Pages
2414 - 2419
Source
2021 IEEE International Conference on Systems, Man, and Cybernetics (SMC)
Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.