Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling

Federated learning that is an approach to addressing the “data silo” problem in a collaborative fashion may face the risk of data leakage in real-world contexts. To solve this problem, we introduce the random Fourier feature mapping (RFFM) together with kernel local differentia...

Full description

Saved in:
Bibliographic Details
Main Authors: Yi Chen, Dan Chen, Niansheng Tang
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10849528/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832576793556549632
author Yi Chen
Dan Chen
Niansheng Tang
author_facet Yi Chen
Dan Chen
Niansheng Tang
author_sort Yi Chen
collection DOAJ
description Federated learning that is an approach to addressing the &#x201C;data silo&#x201D; problem in a collaborative fashion may face the risk of data leakage in real-world contexts. To solve this problem, we introduce the random Fourier feature mapping (RFFM) together with kernel local differential privacy (KLDP) and develop a new privacy protection mechanism, called the RFFM-KLDP mechanism, for high-dimensional context data. Theoretical properties show that the proposed privacy-preserving mechanism has the properties of <inline-formula> <tex-math notation="LaTeX">$\epsilon $ </tex-math></inline-formula>-LDP and <inline-formula> <tex-math notation="LaTeX">$\epsilon $ </tex-math></inline-formula>-distance-LDP in the federated learning framework. To guarantee the effectiveness of federated learning in the presence of contaminated data, we develop a modified low-gradient sampling technique to sample representative subset of uncontaminated data by incorporating large gradients and unbalanced information. By combining RFFM-KLDP and modified low-gradient sampling technique, we develop a novel and robust federated learning method for classification in the presence of the noisy text data, which can preserve data privacy and largely improve the accuracy of classification algorithm compared to the existing classifiers in terms of the area under curve and classification accuracy. Simulation studies and a context example are used to illustrate the proposed methodologies.
format Article
id doaj-art-e9a6e12ca8a24da7bcf625a7edd60cc4
institution Kabale University
issn 2169-3536
language English
publishDate 2025-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj-art-e9a6e12ca8a24da7bcf625a7edd60cc42025-01-31T00:01:25ZengIEEEIEEE Access2169-35362025-01-0113169591697710.1109/ACCESS.2025.353268310849528Federated Learning Based on Kernel Local Differential Privacy and Low Gradient SamplingYi Chen0https://orcid.org/0009-0006-5156-5898Dan Chen1Niansheng Tang2https://orcid.org/0000-0001-7033-3845Yunnan Key Laboratory of Statistical Modeling and Data Analysis, Yunnan University, Kunming, ChinaYunnan Key Laboratory of Statistical Modeling and Data Analysis, Yunnan University, Kunming, ChinaYunnan Key Laboratory of Statistical Modeling and Data Analysis, Yunnan University, Kunming, ChinaFederated learning that is an approach to addressing the &#x201C;data silo&#x201D; problem in a collaborative fashion may face the risk of data leakage in real-world contexts. To solve this problem, we introduce the random Fourier feature mapping (RFFM) together with kernel local differential privacy (KLDP) and develop a new privacy protection mechanism, called the RFFM-KLDP mechanism, for high-dimensional context data. Theoretical properties show that the proposed privacy-preserving mechanism has the properties of <inline-formula> <tex-math notation="LaTeX">$\epsilon $ </tex-math></inline-formula>-LDP and <inline-formula> <tex-math notation="LaTeX">$\epsilon $ </tex-math></inline-formula>-distance-LDP in the federated learning framework. To guarantee the effectiveness of federated learning in the presence of contaminated data, we develop a modified low-gradient sampling technique to sample representative subset of uncontaminated data by incorporating large gradients and unbalanced information. By combining RFFM-KLDP and modified low-gradient sampling technique, we develop a novel and robust federated learning method for classification in the presence of the noisy text data, which can preserve data privacy and largely improve the accuracy of classification algorithm compared to the existing classifiers in terms of the area under curve and classification accuracy. Simulation studies and a context example are used to illustrate the proposed methodologies.https://ieeexplore.ieee.org/document/10849528/Differential privacyfederated learninglow gradient samplingrandom Fourier feature mapping
spellingShingle Yi Chen
Dan Chen
Niansheng Tang
Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling
IEEE Access
Differential privacy
federated learning
low gradient sampling
random Fourier feature mapping
title Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling
title_full Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling
title_fullStr Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling
title_full_unstemmed Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling
title_short Federated Learning Based on Kernel Local Differential Privacy and Low Gradient Sampling
title_sort federated learning based on kernel local differential privacy and low gradient sampling
topic Differential privacy
federated learning
low gradient sampling
random Fourier feature mapping
url https://ieeexplore.ieee.org/document/10849528/
work_keys_str_mv AT yichen federatedlearningbasedonkernellocaldifferentialprivacyandlowgradientsampling
AT danchen federatedlearningbasedonkernellocaldifferentialprivacyandlowgradientsampling
AT nianshengtang federatedlearningbasedonkernellocaldifferentialprivacyandlowgradientsampling