MixNN: Protection of Federated Learning Against Inference Attacks by Mixing Neural Network Layers - Rapports de recherche et Technique de l'Inria Accéder directement au contenu
Rapport (Rapport De Recherche) Année : 2021

MixNN: Protection of Federated Learning Against Inference Attacks by Mixing Neural Network Layers

Résumé

Machine Learning (ML) has emerged as a core technology to provide learning models to perform complex tasks. Boosted by Machine Learning as a Service (MLaaS), the number of applications relying on ML capabilities is ever increasing. However, ML models are the source of different privacy violations through passive or active attacks from different entities. In this paper, we present MixNN a proxy-based privacy-preserving system for federated learning to protect the privacy of participants against a curious or malicious aggregation server trying to infer sensitive information (i.e., membership and attribute inferences). MixNN receives the model updates from participants and mixes layers between participants before sending the mixed updates to the aggregation server. This mixing strategy drastically reduces privacy leaks without any trade-off with utility. Indeed, mixing the updates of the model has no impact on the result of the aggregation of the updates computed by the server. We report on an extensive evaluation of MixNN using several datasets and neural networks architectures to quantify privacy leakage through membership and attribute inference attacks as well the robustness of the protection. We show that MixNN significantly limits both the membership and attribute inferences compared to a baseline using model compression and noisy gradient (well known to damage the utility) while keeping the same level of utility as classic federated learning.
Fichier principal
Vignette du fichier
MixNN (19).pdf (1.82 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03354724 , version 1 (26-09-2021)

Identifiants

  • HAL Id : hal-03354724 , version 1

Citer

Antoine Boutet, Thomas Lebrun, Jan Aalmoes, Adrien Baud. MixNN: Protection of Federated Learning Against Inference Attacks by Mixing Neural Network Layers. [Research Report] RR-9411, INRIA Grenoble. 2021, pp.1-21. ⟨hal-03354724⟩
81 Consultations
116 Téléchargements

Partager

Gmail Facebook X LinkedIn More