Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling
Federated learning that is an approach to addressing the “data silo” problem in a collaborative fashion may face the risk of data leakage in real-world contexts. To solve this problem, we introduce the random Fourier feature mapping (RFFM) together with kernel local differentia...
Saved in:
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2025-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10849528/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1832576793556549632 |
---|---|
author | Yi Chen Dan Chen Niansheng Tang |
author_facet | Yi Chen Dan Chen Niansheng Tang |
author_sort | Yi Chen |
collection | DOAJ |
description | Federated learning that is an approach to addressing the “data silo” problem in a collaborative fashion may face the risk of data leakage in real-world contexts. To solve this problem, we introduce the random Fourier feature mapping (RFFM) together with kernel local differential privacy (KLDP) and develop a new privacy protection mechanism, called the RFFM-KLDP mechanism, for high-dimensional context data. Theoretical properties show that the proposed privacy-preserving mechanism has the properties of <inline-formula> <tex-math notation="LaTeX">$\epsilon $ </tex-math></inline-formula>-LDP and <inline-formula> <tex-math notation="LaTeX">$\epsilon $ </tex-math></inline-formula>-distance-LDP in the federated learning framework. To guarantee the effectiveness of federated learning in the presence of contaminated data, we develop a modified low-gradient sampling technique to sample representative subset of uncontaminated data by incorporating large gradients and unbalanced information. By combining RFFM-KLDP and modified low-gradient sampling technique, we develop a novel and robust federated learning method for classification in the presence of the noisy text data, which can preserve data privacy and largely improve the accuracy of classification algorithm compared to the existing classifiers in terms of the area under curve and classification accuracy. Simulation studies and a context example are used to illustrate the proposed methodologies. |
format | Article |
id | doaj-art-e9a6e12ca8a24da7bcf625a7edd60cc4 |
institution | Kabale University |
issn | 2169-3536 |
language | English |
publishDate | 2025-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Access |
spelling | doaj-art-e9a6e12ca8a24da7bcf625a7edd60cc42025-01-31T00:01:25ZengIEEEIEEE Access2169-35362025-01-0113169591697710.1109/ACCESS.2025.353268310849528Federated Learning Based on Kernel Local Differential Privacy and Low Gradient SamplingYi Chen0https://orcid.org/0009-0006-5156-5898Dan Chen1Niansheng Tang2https://orcid.org/0000-0001-7033-3845Yunnan Key Laboratory of Statistical Modeling and Data Analysis, Yunnan University, Kunming, ChinaYunnan Key Laboratory of Statistical Modeling and Data Analysis, Yunnan University, Kunming, ChinaYunnan Key Laboratory of Statistical Modeling and Data Analysis, Yunnan University, Kunming, ChinaFederated learning that is an approach to addressing the “data silo” problem in a collaborative fashion may face the risk of data leakage in real-world contexts. To solve this problem, we introduce the random Fourier feature mapping (RFFM) together with kernel local differential privacy (KLDP) and develop a new privacy protection mechanism, called the RFFM-KLDP mechanism, for high-dimensional context data. Theoretical properties show that the proposed privacy-preserving mechanism has the properties of <inline-formula> <tex-math notation="LaTeX">$\epsilon $ </tex-math></inline-formula>-LDP and <inline-formula> <tex-math notation="LaTeX">$\epsilon $ </tex-math></inline-formula>-distance-LDP in the federated learning framework. To guarantee the effectiveness of federated learning in the presence of contaminated data, we develop a modified low-gradient sampling technique to sample representative subset of uncontaminated data by incorporating large gradients and unbalanced information. By combining RFFM-KLDP and modified low-gradient sampling technique, we develop a novel and robust federated learning method for classification in the presence of the noisy text data, which can preserve data privacy and largely improve the accuracy of classification algorithm compared to the existing classifiers in terms of the area under curve and classification accuracy. Simulation studies and a context example are used to illustrate the proposed methodologies.https://ieeexplore.ieee.org/document/10849528/Differential privacyfederated learninglow gradient samplingrandom Fourier feature mapping |
spellingShingle | Yi Chen Dan Chen Niansheng Tang Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling IEEE Access Differential privacy federated learning low gradient sampling random Fourier feature mapping |
title | Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling |
title_full | Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling |
title_fullStr | Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling |
title_full_unstemmed | Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling |
title_short | Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling |
title_sort | federated learning based on kernel local differential privacy and low gradient sampling |
topic | Differential privacy federated learning low gradient sampling random Fourier feature mapping |
url | https://ieeexplore.ieee.org/document/10849528/ |
work_keys_str_mv | AT yichen federatedlearningbasedonkernellocaldifferentialprivacyandlowgradientsampling AT danchen federatedlearningbasedonkernellocaldifferentialprivacyandlowgradientsampling AT nianshengtang federatedlearningbasedonkernellocaldifferentialprivacyandlowgradientsampling |