Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

nips23-Dynamic Personalized Federated Learning with Adaptive Differential Privacy

Notifications You must be signed in to change notification settings

lq1029/DynamicPFL

Folders and files

NameName
Last commit message
Last commit date

Latest commit

History

12 Commits

Repository files navigation

Dynamic Personalized Federated Learning with Adaptive Differential Privacy

Dynamic Personalized Federated Learning with Adaptive Differential Privacy,

Xiyuan Yang, Wenke Huang,Mang Ye NeurIPS, 2023 Link

News

  • [2023年10月08日] Code has released.

Abstract

Personalized federated learning with differential privacy (DP-PFL) has been considered a feasible solution to address non-IID distribution of data and privacy leakage risks. However, current DP-PFL methods suffer from inflexible personalization and convergence difficulties due to two main factors: 1) Firstly, we observe that the prevailing personalization methods mainly achieve this by personalizing a fixed portion of the model, which lacks flexibility. 2) Moreover, we further demonstrate that the default gradient calculation is sensitive to the widely-used clipping operations in differential privacy, resulting in difficulties in convergence. Considering that Fisher information values can serve as an effective measure for estimating the information content of parameters by reflecting the model sensitivity to parameters, we aim to leverage this property to address the aforementioned challenges. In this paper, we propose a novel adaptive method for DP-PFL to handle these challenges. Firstly, by using layer-wise Fisher information to measure the information content of local parameters, we retain local parameters with high Fisher values during the personalization process, which are considered informative, simultaneously prevent these parameters from noise perturbation. Secondly, we introduce an adaptive approach by applying differential constraint strategies to personalized parameters and shared parameters identified in the previous for better convergence. Our method boosts performance through flexible personalization while mitigating the slow convergence caused by clipping operations. Experimental results on CIFAR-10, FEMNIST and SVHN dataset demonstrate the effectiveness of our approach in achieving better personalization performance and robustness against clipping, under differential privacy personalized federated learning.

Citation

@inproceedings{nips23dynamicPFL,
title={Dynamic Personalized Federated Learning with Adaptive Differential Privacy},
author={Yang, Xiyuan and Huang, Wenke and Ye, Mang},
booktitle={Thirty-seventh Conference on Neural Information Processing Systems},
year={2023},
}

example usage

python main_base.py
python ours.py

Relevant Projects

[1] Rethinking Federated Learning with Domain Shift: A Prototype View - CVPR 2023 [Link]

About

nips23-Dynamic Personalized Federated Learning with Adaptive Differential Privacy

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%

AltStyle によって変換されたページ (->オリジナル) /