Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2025 Jul 22;15:26679. doi: 10.1038/s41598-025-12575-6

ALDP-FL for adaptive local differential privacy in federated learning

Lixin Cui 1, Xu Wu 1,
PMCID: PMC12284227  PMID: 40696085

Abstract

Federated learning, as an emerging distributed learning framework, enables model training without compromising user data privacy. However, malicious attackers may still infer sensitive user information by analyzing model updates during the federated learning process. To address this, this paper proposes an Adaptive Localized Differential Privacy Federated Learning (ALDP-FL) method. This approach dynamically sets the clipping threshold for each network layer's updates based on the historical moving average of their Inline graphic-norm, thereby injecting adaptive noise into each layer. Additionally, a bounded perturbation mechanism is designed to minimize the impact of the added noise on model accuracy. A privacy analysis of the method is provided. Finally, experiments on the MNIST, Fashion MNIST, and CIFAR-10 datasets demonstrate the effectiveness and practicality of the proposed method. Specifically, ALDP-FL achieves an average improvement of over 10% across all evaluation metrics: Accuracy increases by 10.57%, Precision by 10.64%, Recall by 10.52%, and F1 Score by 10.64%. Regarding the reconstructed images under the iDLG attack, the average improvement rates in MSE and SSIM reach 391.2% and –85.4%, respectively, significantly outperforming all other comparison methods.

Subject terms: Mathematics and computing, Computer science, Information technology

Introduction

In today's data-driven era, the tension between privacy protection and data sharing becomes increasingly prominent1. Many traditional machine learning approaches2 rely on centralized data storage and processing, which can enhance model performance3. However, in sensitive fields such as healthcare, finance, and the Internet of Things, data often exhibits data silos due to privacy regulations or commercial confidentiality, making cross-institutional sharing difficult4. Consequently, how to leverage massive datasets for model training while safeguarding user privacy has become a significant research challenge.5

Federated Learning (FL)6, as an emerging distributed machine learning framework, enables model training across multiple user devices without centralizing data, effectively protecting user privacy79. Nonetheless, studies show that despite FL providing a layer of privacy protection10, it still faces threats such as model inversion attacks, membership inference attacks, and gradient leakage. Attackers can analyze intermediate parameters to infer sensitive information from original training data, making privacy preservation a critical bottleneck for deploying FL systems11,12. Yin13 et al. categorize potential privacy leakage risks into five fundamental aspects: attacker, attack method, attack timing, attack location, and attack motivation. Gupta14 et al. proposed a data poisoning attack, which reverses the benign model's loss function to degrade the global model's performance. Gu15 et al. introduced CS-MIA, a novel membership inference attack based on prediction confidence sequences, posing serious privacy threats to FL. Zhu16 et al. demonstrated that private data could be obtained from publicly shared gradients, referring to this type of privacy breach as gradient deep leakage.

Currently, privacy enhancement techniques for FL primarily fall into three categories: Differential Privacy17 (DP), Homomorphic Encryption, and Secure Multi-Party Computation18. Differential privacy adds random noise to data or model outputs to hide individual information19. However, its noise injection often slows convergence and reduces accuracy, especially under non-i.i.d. data distributions20, making it challenging to balance privacy budgets and model utility. Homomorphic encryption and secure multi-party computation provide strong cryptographic guarantees but incur high computational and communication overhead, limiting their applicability to large-scale deep neural network training21.

This paper mainly investigates the application of differential privacy to address privacy and security issues in federated learning. Existing research includes McMahan22 et al., who first applied differential privacy in FL by adding Gaussian noise to the aggregated model at the central server after weighted parameter aggregation. However, they did not consider whether the central server is honest or trustworthy. Yang23 et al. proposed GFL-ALDPA, a gradient-compressed federated learning framework based on adaptive local differential privacy budget allocation. They developed a novel adaptive privacy budget allocation scheme based on communication rounds to reduce privacy budget wastage and model noise. By allocating different privacy budgets to distinct communication rounds during training, this approach maximizes the utilization of limited privacy budgets while enhancing model accuracy. Moreover, they introduced a dimension reduction-based gradient compression mechanism that simultaneously reduces communication costs, overall noise magnitude, and total privacy budget consumption of the model, thereby ensuring accuracy under specified privacy protection guarantees. Truex24 et al. developed the LDP-Fed local differential privacy method, which introduces privacy protection during local training by perturbing gradients, achieving localized differential privacy. They employed fixed gradient clipping and fixed noise addition. Yang25 et al. introduced Fed-DPA, which uses a layer-wise Fisher information-based dynamic personalized strategy to achieve flexible personalization while reducing noise impact. Most current research focuses on static defenses against single attack scenarios, lacking systematic protection against multi-layer attack chains in dynamic adversarial environments. This fragmentation results in existing solutions facing dilemmas of either over-protection leading to performance loss or incomplete defenses leaving vulnerabilities in actual deployment.

To address the aforementioned challenges, this paper proposes a novel adaptive local differential privacy federated learning method. This method innovatively employs a sliding window mechanism to introduce adaptive noise for clients. The incorporation of the sliding window mechanism not only maintains model training accuracy but also integrates adaptive differential privacy protection into the data, effectively addressing the limitations of existing approaches.

Furthermore, to mitigate the impact of differential privacy on model performance, this study designs a bounded perturbation mechanism. A bounded perturbation function Inline graphic, composed of the variance of added noise, the privacy budget of differential privacy, and a bounding factor, is utilized to constrain the range of noise injected into the data. This function organically combines these parameters to enhance the efficiency of differential privacy protection.We conducted three sets of experiments on the MNIST and Fashion MNIST datasets to validate the effectiveness and practicality of the ALDP-FL method.

In summary, this paper makes the following contributions:

  1. Proposes an Adaptive Local Differential Privacy Federated Learning (ALDP-FL) method, which dynamically sets the clipping threshold for each network layer based on the historical moving average of the Inline graphic norms of layer updates, adding adaptive noise accordingly.

  2. Designs a bounded perturbation mechanism that constrains the noise range to reduce the accuracy loss of the federated model caused by local differential privacy.

  3. Through extensive experiments on MNIST, Fashion MNIST and CIFAR-10 datasets, we demonstrate the effectiveness, scalability, and practicality of our method. Results show significant improvements in privacy-utility trade-offs compared to state-of-the-art baselines.

Related work

Federated learning acts as a distributed machine learning framework, where models are trained locally and only model parameters are uploaded to the server, significantly reducing the risks of data transmission and leakage. However, despite this, federated learning still faces various privacy and security challenges, such as model inversion, membership inference, and data poisoning26. Li27 et al. categorized existing research on model inversion attacks in FL into two types: update-based attacks, which utilize one or more historical versions of the target model to infer membership information; and trend-based attacks, which analyze the trajectory of specific metrics to determine membership status. Zhao28 et al. discovered that sharing gradients indeed leaks the true labels, and they proposed a simple yet reliable method to extract accurate data from the gradients. Their method can reliably extract the true labels, and they named it the improved DLG (iDLG). Their method is applicable to any differentiable model trained with cross-entropy loss on one-hot labels. They mathematically explained how their method extracts the true labels from gradients and demonstrated its advantages over DLG through experiments. Khraisat29 et al. investigated targeted data poisoning attacks in FL systems, where a small fraction of malicious participants compromise the global model through updates with mislabeled data. Their findings reveal that even a limited number of malicious participants significantly degrades classification accuracy and recall, particularly when attacks concentrate on specific classes. They also examined both the duration and timing of these attacks during early and late training rounds, highlighting the impact of malicious participant availability on attack effectiveness.

The above studies demonstrate that federated learning has many security issues. Researchers have proposed some corresponding solutions. Depending on different attack types, they suggest techniques30 such as encryption, differential privacy, and anomaly detection to safeguard the security of federated learning. Tripathy31 et al. proposed the HalfFedLearn framework, which uses homomorphic encryption and local horizontal data partitioning to address challenges such as slow convergence, high computational and communication costs, and the security of parameter sharing. They leveraged the inherent distribution of the dataset, employed horizontal data partitioning based on data sensitivity, and implemented selective privacy protection for private data samples using homomorphic encryption. While homomorphic encryption provides robust protection for data privacy, it also consumes significant computational resources when performing operations on encrypted data32. The volume of encrypted data expands substantially, leading to increased bandwidth requirements for transmission and higher storage costs33. This challenge is particularly pronounced in federated learning scenarios, which require frequent exchanges of gradients or model updates34. Furthermore, in multi-party settings, strict security mechanisms are essential for key generation, distribution, updating, and revocation35. If keys are compromised, the security of all encrypted data would be jeopardized, posing additional challenges to federated learning systems.

The use of differential privacy (DP) technology to protect privacy in federated learning36 addresses a series of issues caused by homomorphic encryption (HE). By adding controlled noise (e.g., Laplace noise, Gaussian noise)37 during the local training or parameter aggregation phase, differential privacy prevents attackers from inferring individual data through gradient analysis. Xie38 et al. proposed a new privacy-preserving federated learning framework, enhanced by adaptive differential privacy, to achieve secure medical data collaboration. They introduced a dual-layer privacy protection mechanism combining local and central differential privacy, dynamically adjusting the allocation of privacy budgets based on training progress and data sensitivity. The framework implements a hierarchical architecture where edge servers perform preliminary aggregation to reduce communication overhead and enhance privacy protection. A novel adaptive privacy budget allocation strategy was developed to optimize the privacy-utility trade-off throughout the training process.

While differential privacy enhances data privacy in federated learning by adding noise, this approach introduces bias, particularly when data is sparse or the privacy budget is small39. Such bias can slow down model convergence or significantly degrade final accuracy. Additionally, to prevent noise amplification, gradient norm clipping40 must be applied. However, selecting appropriate clipping thresholds requires repeated experimentation and may introduce clipping bias41. Furthermore, in multi-round training, the privacy budget must be dynamically allocated for each round to avoid exceeding cumulative privacy leakage thresholds42. These challenges highlight the complexity of jointly optimizing and tuning parameters such as43 privacy budgets, noise distributions, and clipping thresholds, which often leads to prohibitive debugging costs in practical implementations.Wang44 et al. proposed a novel adaptive differential privacy method with feedback regulation, called FedFR-ADP. First, they employed Earth Mover's Distance (EMD) to measure the data heterogeneity of each client and adaptively applied Gaussian noise based on the degree of heterogeneity, making the noise addition more targeted and effective. Second, they introduced a feedback regulation mechanism to dynamically adjust the privacy budget based on the error feedback of the global model, further enhancing the model's performance. Talaei45 et al. proposed injecting adaptive noise into FL based on the relative importance of data features. They first introduced two effective methods for prioritizing features in deep neural network models and then perturbed the model weights based on this information. Specifically, they explored whether the idea of adding more noise to less important parameters and less noise to more important parameters could effectively preserve model accuracy while protecting privacy. Their experiments confirmed this claim under certain conditions. The amount of noise injected, the proportion of parameters involved, and the number of global iterations can significantly alter the output. While carefully selecting parameters by considering dataset attributes can enhance privacy without severely compromising accuracy, incorrect choices may degrade model performance.

Despite significant advancements in existing research addressing privacy and security challenges in federated learning, current frameworks still exhibit notable limitations. For instance, homomorphic encryption-based approaches suffer from computational resource inefficiency, excessive communication overhead, and risks of key leakage. Meanwhile, differential privacy-based methods face challenges in balancing privacy budget allocation, setting appropriate clipping thresholds, and optimizing noise distributions. This study aims to enhance privacy protection in federated learning through differential privacy techniques. Specifically, we propose an adaptive clipping threshold mechanism to dynamically adjust noise injection, thereby minimizing the adverse impact of noise on model performance while ensuring robust data privacy safeguards.

Overview of federated learning

Federated learning is a privacy-preserving distributed collaborative learning framework46, primarily aimed at achieving joint model training across multiple participants without sharing raw data, as shown in Fig. 1 of its framework.

Fig. 1.

Fig. 1

Federated learning framework.

This approach employs decentralized local computation and parameter aggregation mechanisms to effectively mitigate data silo issues while ensuring privacy compliance47. Assuming the system involves Inline graphic participants, with the Inline graphic-th participant possessing a private dataset Inline graphic of size Inline graphic, and the total global sample size being Inline graphic. The optimization objective of federated learning48 is to minimize the weighted global empirical risk:

graphic file with name d33e435.gif 1

Among them, Inline graphic represents the empirical risk function for the Inline graphic-th local dataset, where Inline graphic denotes the local loss function for data point Inline graphic, the P subset involves participant data with weighted contributions.

The federated learning process can be divided into three phases. Local model updates49: Each participant performs model updates based on their local private data, starting from the current global model, and executes gradient descent steps. Parameter aggregation: After the participants' updates are uploaded to the server, the server executes the model aggregation algorithm, merging the updates from each participant into a global model update. This process yields a new global model. Model synchronization and iteration: After updating, the global model parameters are broadcast back to all participants. This process repeats until convergence criteria are met, such as reaching a predefined number of communication rounds or achieving desired accuracy.

Local differential privacy

Local Differential Privacy50 (LDP) is a distributed privacy-preserving model that fundamentally involves users directly perturbing sensitive data on their local devices before submitting it to a server for statistical analysis or model training. Unlike centralized differential privacy, where data is collected first and then anonymized, LDP ensures that the privacy of individual data points is protected even before data transmission. It is particularly suitable for IoT, federated learning, and various distributed scenarios.

The Gaussian Mechanism51, as an example of an LDP method, adds Gaussian noise to continuous data to achieve privacy guarantees while maintaining data utility. This mechanism supports an (Inline graphic)-differential privacy definition, ensuring that for all neighboring datasets Inline graphic differing in one element, and for all possible output sets Inline graphic, the following holds:

graphic file with name d33e505.gif 2

Among them, Inline graphic represents the perturbation mechanism, Inline graphic is the privacy budget, and Inline graphic is the relaxation parameter. For a user with local data Inline graphic, the released output after perturbation is modeled as Inline graphic, where Inline graphic is a multivariate Gaussian noise with mean zero and covariance matrix Inline graphic. The noise scale Inline graphic is typically determined based on the privacy parameters Inline graphic and the sensitivity Inline graphic, which is set to satisfy the Inline graphic, where Inline graphic is the Inline graphic-dimensional Inline graphic sensitivity, and Inline graphic is the sensitivity of the function under consideration.

Methods

The overall workflow of the ALDP-FL framework

Federated learning, as a distributed machine learning framework, achieves privacy protection through client-side local training and parameter aggregation. However, privacy security issues still persist. The locally updated parameters inherently encode features of the raw data, and attackers can potentially reconstruct users' training samples by analyzing gradient update information. Traditional federated learning architectures assume the central server is honest and trustworthy, but if the central server is curious, clients may inadvertently leak their private data during collaboration.

Algorithm 1.

Algorithm 1

ALDP-FL

To address these issues, this paper proposes an adaptive local differential privacy-based federated learning method. This approach ensures that clients can safely and effectively participate in federated model training with an honest-but-curious central server. Even if an attacker intercepts the transmitted model parameters, the privacy of local data can still be preserved.

The ALDP-FL algorithm is shown as Algorithm 1, The relevant symbols and parameters52 involved in this paper are as shown in Table 1. And its workflow is as follows: First, initialize the global model parameters as Inline graphic, then perform T rounds of iterative training. In each round, the server selects K clients from N clients as participants for that round and distributes the current global model Inline graphic to these participants. After local training, each participant returns its model update Inline graphic to the server. The server aggregates all participants' updates by computing the arithmetic mean Inline graphic to obtain the global model update for this round, and correspondingly updates the global model as Inline graphic. This process repeats until T rounds are completed.

Table 1.

Relevant symbols and parameters.

Symbol Configuration
N Total number of clients
K Number of participants per round
T Total number of global iterations
Inline graphic Global model at round t
E Client iteration round
D Local training dataset
Inline graphic Local initialization model at round e
Inline graphic The local model updated in round e (without differential privacy protection)
Inline graphic Update amount of the local model in round e
H A sliding window storing local model updates
Inline graphic Clipping factor for round e
Inline graphic Update after clipping in round e
Inline graphic Sensitivity
k Number of sliding windows
Inline graphic Update with differential privacy in round e
Inline graphic Model update of the k-th participant with differential privacy protection
B Bounded perturbation function
Inline graphic Gaussian distribution standard deviation
Inline graphic Parameters related to DP
Inline graphic Hyperparameters in bounded perturbation functions

The ALDP-FL method framework primarily consists of two parts: the server-side algorithm and the client-side algorithm, with the specific workflow illustrated in Fig. 2. The server-side algorithm is similar to traditional federated learning server algorithms, responsible for initializing model parameters, aggregating data from participating clients, and updating the global model. The client-side algorithm employs an adaptive local differential privacy mechanism designed to ensure the privacy and security of client data.

Fig. 2.

Fig. 2

ALDP-FL framework process.

The core concept of the adaptive local differential privacy algorithm revolves around adaptivity and localization. Localization refers to performing all differential privacy-related processes locally on the client side, while adaptivity means that the noise perturbation introduced in differential privacy is dynamically adjusted based on the current state of the model. Pruning is a critical technique in differential privacy; choosing the appropriate pruning threshold for protected data is particularly important. Fixed pruning thresholds may lead to excessive information loss and impair model accuracy if set too low, or insufficient privacy guarantees if set too high. The adaptive local differential privacy mechanism dynamically adjusts pruning thresholds for different neural network layers during each iteration, thereby addressing the limitations of fixed thresholds.

To select the pruning threshold for each neural network layer at each iteration, a temporal moving average approach is employed, leveraging the temporal similarity principle. This principle states that system behavior at nearby time points tends to be similar. By using the average change in previous rounds as the current threshold, the method effectively reduces information loss caused by pruning and enables the threshold to be adaptively updated. After determining the pruning threshold, Gaussian noise is added to the updates by incorporating bounded perturbation mechanisms, providing differential privacy guarantees. This not only significantly enhances data privacy but also helps maintain model accuracy.

ALDP-FL client-side algorithm

The client-side algorithm is shown in Algorithm 2, and the flowchart is depicted in Fig. 3. After receiving the model Inline graphic sent from the central server, the client initializes the local model Inline graphic, copying the global model parameters Inline graphic into the local model Inline graphic. It then performs E rounds of local training loops, where each round represents a complete local training iteration. In each iteration, gradient descent updates are first applied to each batch d from the local dataset D as follows:Inline graphic where Inline graphic is the learning rate and Inline graphic is the gradient of the loss function L with respect to the model parameters Inline graphic. Based on the gradient descent update, the parameter update for the current iteration is computed as Inline graphic. The Inline graphic norm of Inline graphic for the current iteration is then calculated and stored in a sliding window H.

Fig. 3.

Fig. 3

Flowchart of the ADP-FL method client-side process.

The update for each model layer, Inline graphic, is clipped using Inline graphic to obtain the clipped update Inline graphic. Inline graphic is a dynamic value, which is calculated based on the average Inline graphic norm of the most recent k updates (where k is the sliding window size), as shown in Fig. 4. When Inline graphic, the update Inline graphic remains unchanged. Otherwise, Inline graphic is clipped to Inline graphic.

Algorithm 2.

Algorithm 2

Client-side

Fig. 4.

Fig. 4

The process of calculating Inline graphic.

Based on the local differential privacy definition, in order to ensure that the participants satisfy Inline graphic-differential privacy during local training, Gaussian noise must be added to the clipped update Inline graphic. The standard deviation of the noise is:Inline graphicwhere Inline graphic is the slack parameter of differential privacy and Inline graphic is the privacy budget. Based on the noise standard deviation and the bounded perturbation function B, the differential privacy-protected update is obtained by adding noise to the clipped update Inline graphic as follows:Inline graphic Finally, the model update for the current round is computed as:Inline graphic. This process repeats for E rounds to complete the local model training. After completing E rounds of local training, the participant's update is calculated as:Inline graphicand this value is returned to the server.

Regarding the Inline graphic value proposed above, a mathematical proof demonstrating that this value ensures Inline graphic satisfies Inline graphic-differential privacy will be provided in the following sections. Additionally, the bounded perturbation mechanism B mentioned earlier will be concretely designed and analyzed in subsequent content.

Bounded perturbation mechanisms

The ALDP-FL method proposed in this paper selects the Gaussian distribution as the source of noise for differential privacy. The probability density function of the Gaussian distribution is a symmetric bell-shaped curve. The curve is unimodal, reaching its highest point at the mean and is perfectly symmetric about the mean, meaning the distribution shapes on the left and right sides of the mean are identical. When the mean of the Gaussian distribution is set to 0, zero-mean noise can be added, avoiding systematic bias. However, since the Gaussian distribution has infinite support, its long-tail effect may potentially harm the model's performance. This paper designs a bounded perturbation mechanism to eliminate the possible impact of the long-tail effect of the Gaussian distribution. The added noise is constrained within the bounds of [–b, b].

graphic file with name d33e1169.gif 3

Among them, Inline graphic is the privacy-preserving privacy budget, Inline graphic is a superparameter, and Inline graphic.

graphic file with name d33e1194.gif 4

The function Inline graphic represents a high-dimensional probability density function with a mean value of 0 and a standard deviation Inline graphic. Since we are limited to the range of the high-dimensional noise values, we need to normalize the high-dimensional distribution. The normalized high-dimensional probability density function after normalization is defined as Inline graphic. Here, Inline graphic is a normalization constant.

graphic file with name d33e1226.gif 5
graphic file with name d33e1232.gif 6

The cumulative distribution function of the standard normal distribution is introduced into the normalization constant.

graphic file with name d33e1240.gif 7

The standard deviation Inline graphic of the truncated normal distribution can be used to estimate the standard deviation Inline graphic.

graphic file with name d33e1259.gif 8

According to the standard deviation of the normal distribution, we can use the Hoeffding inequality53 to calculate the deviation of the empirical mean from the true mean of the distribution. Hoeffding inequality is a type of concentration inequality that does not require the assumption of independence and identical distribution of random variables. It estimates the probability that the sample mean deviates from the true mean by a certain amount. Suppose Inline graphic are independent random variables bounded between Inline graphic and Inline graphic, and their sample mean is Inline graphic. Then, for any Inline graphic, Hoeffding's inequality can be expressed as:

graphic file with name d33e1301.gif 9

Here, Inline graphic is the expectation of the random variable Inline graphic, and Inline graphic is the error bound for the mean estimation.

Since the mean of the truncated Gaussian distribution is zero, Inline graphic. Substituting the range of the truncated Gaussian distribution (-b, b) into Hoeffding's inequality, we obtain:

graphic file with name d33e1335.gif 10

This inequality indicates that, with a fixed probability Inline graphic, the sample mean Inline graphic deviates from the true mean 0 with a probability not exceeding Inline graphic. For convenience, we can set the confidence level to Inline graphic, and solve for the corresponding Inline graphic.

graphic file with name d33e1373.gif 11

From the above analysis, it can be concluded that the noise processed by the boundary perturbation mechanism satisfies the high-dimensional Gaussian distribution Inline graphic.Moreover, with at least a probability of Inline graphic, the average noise value Inline graphic falls within the interval Inline graphic. Although this method improves the accuracy of the model, it reduces the precision of noise sampling.

Privacy-security analysis

To ensure that the ADP-FL presented in this paper satisfies Inline graphic-privacy, we need to demonstrate that for the corresponding datasets Inline graphic and Inline graphic, as well as any output set Inline graphic, the following condition holds:

graphic file with name d33e1433.gif 12

Suppose Inline graphic and Inline graphic are the outputs of the function Inline graphic on neighboring datasets Inline graphic and Inline graphic. As can be seen from the above text, the sensitivity Inline graphic. The output of the Gaussian mechanism is:

graphic file with name d33e1478.gif 13
graphic file with name d33e1484.gif 14

where Inline graphic is a random variable drawn from a Gaussian distribution with mean zero and variance Inline graphic. Therefore, we need to prove that:

graphic file with name d33e1503.gif 15

Calculate the ratio of probabilities of Inline graphic and Inline graphic:

graphic file with name d33e1523.gif 16

For any chosen Inline graphic, assuming Inline graphic lies within the interval Inline graphic, perform boundary analysis on Inline graphic.

graphic file with name d33e1555.gif 17

According to the definition of differential privacy, we need to satisfy:

graphic file with name d33e1562.gif 18

Certainly:

graphic file with name d33e1569.gif 19

To consider the impact of Inline graphic, we need to use the properties of the Gaussian mechanism:

graphic file with name d33e1582.gif 20

Finally, the lower bound of Inline graphic is:

graphic file with name d33e1596.gif 21

Therefore, in each addition of Gaussian noise, choose Inline graphic, so that the participant's local training process satisfies the differential privacy Inline graphic.

Experiment

Experimental setup

We employed two convolutional neural network models, LeNet and ResNet-18, to test the ALDP-FL method. LeNet consists of a total of 7 layers, including 2 convolutional layers, 2 pooling layers, and 3 fully connected layers. It is a lightweight and simple model, making it suitable for relatively straightforward tasks such as basic image classification. ResNet-18 consists of 18 layers, including 17 convolutional layers and 1 fully connected layer. It is commonly used for more complex image classification tasks due to its deep architecture and ability to handle more intricate patterns.

We utilized three datasets: MNIST, Fashion MNIST, and CIFAR-10. The MNIST dataset contains 70,000 grayscale images of handwritten digits, each 28x28 pixels, with 60,000 images for training and 10,000 for testing. The labels range from 0 to 9, each corresponding to a handwritten digit. The Fashion MNIST dataset also consists of 70,000 grayscale images with the same dimensions as MNIST, with 60,000 for training and 10,000 for testing. However, these images represent 10 different clothing categories, such as T-shirt, Trouser, and Pullover, making it more complex than MNIST. The CIFAR-10 dataset consists of 60,000 color images, each 32x32 pixels with 3 color channels, of which 50,000 are used for training and 10,000 for testing. These images belong to 10 common object categories, such as Airplane, Automobile, and Bird, making the CIFAR-10 dataset more challenging for method evaluation. The hardware platform and environmental parameters used during the experimental training process are shown in Table 2.

Table 2.

The hardware platform and environmental parameters.

Parameters Configuration
CPU Intel(R) Xeon(R) Gold 6226R
Memory 64 GB
Operating System Windows 11
GPU NVIDIA GeForce RTX 3090
Deep Learning Framework pytorch

In our experiments, we set the total number of FL clients, Inline graphic, and the number of participants per global iteration, Inline graphic. We set the total number of global iterations, Inline graphic, and the number of local iterations per participant, Inline graphic. The batch size was set to 128, and the learning rate Inline graphic was 0.001. For the optimization algorithm, we employed the Adam optimizer, known for its adaptive learning rate capabilities. For the incorporated differential privacy, we set the privacy budget per client, Inline graphic, and the slack variable Inline graphic.

To verify the effectiveness of the proposed ALDP-FL method, we compared it with existing methods: FedAvg (without differential privacy protection), Fed-DPA, GFL-ALDPA, and LDP-Fed. We designed three sets of experiments:

  1. Using different models and datasets to validate the impact of the sliding window on performance.

  2. Comparing the performance of ALDP-FL with state-of-the-art methods across different models and datasets.

  3. Testing the privacy protection capabilities of different methods through model inversion attacks.

Evaluation metrics

To conduct a quantitative evaluation of ALDP-FL's comprehensive performance, we employed four widely adopted metrics: Accuracy, the most intuitive metric, represents the proportion of correct predictions; Precision measures the accuracy of positive predictions, Recall evaluates the ability to identify positive samples, and F1-score balances both precision and recall.

For evaluating ALDP-FL's privacy-preserving capabilities, we used four established metrics: Mean Squared Error (MSE) calculates the average squared difference between pixel values, with lower values indicating superior reconstruction quality; Peak Signal-to-Noise Ratio (PSNR) measures the ratio of signal to distortion power, being sensitive to uniform distortions but less so to structural ones; Mean Absolute Error (MAE) computes the average absolute difference between pixels, providing an intuitive interpretation of per-pixel error; and Structural Similarity Index (SSIM) assesses luminance, contrast, and structural similarity on a [0,1] scale, with higher values denoting greater similarity and better alignment with human visual perception.

Experimental validation of the impact of sliding window on ALDP-FL performance

This set of experiments aims to explore the optimal sliding window size k for the ALDP-FL method when the local iteration rounds E is 10, under different models and datasets, to achieve the best performance of the global model. We set the range of the sliding window k to [2,5].

The experimental process is illustrated in Fig. 5, where the first row presents the results obtained using the LeNet-5 architecture, and the second row corresponds to ResNet-18. On the MNIST dataset, all configurations converge rapidly within the initial training epochs, achieving accuracies exceeding 90%. As shown in Fig. 5a (top row), LeNet-5 achieves optimal performance when the sliding window size is Inline graphic, marginally outperforming other settings. A larger window size (Inline graphic) results in a slight decline in accuracy, likely due to the simplicity of the dataset, where excessive contextual information may introduce redundancy rather than benefit learning. In the case of ResNet-18 (Fig. 5a, bottom row), the performance gap across different k values is minimal, though Inline graphic and Inline graphic still yield marginally higher final accuracies, suggesting that moderate window sizes contribute positively even in deeper models.

Fig. 5.

Fig. 5

Comparison test of different sliding windows.

For the Fashion MNIST dataset (Fig. 5b), which presents greater intra-class variability and visual complexity than MNIST, the influence of window size becomes more pronounced. With LeNet-5, increasing k from 2 to 3 leads to a noticeable improvement in performance; however, further enlargement to Inline graphic or Inline graphic does not result in continued gains and even causes a minor accuracy drop, especially at early stages of training. Notably, Inline graphic produces the lowest accuracy during the initial epochs. This observation suggests that excessive window size may introduce noise and hinder generalization in shallow networks. ResNet-18 demonstrates improved robustness across varying window sizes due to its enhanced representational capacity and residual connections, though Inline graphic remains the most effective setting.

The CIFAR-10 dataset, characterized by complex spatial features and significant inter-class variation, further amplifies the impact of network depth and window configuration (Fig. 5c). LeNet-5 exhibits limited performance, failing to exceed 70% accuracy under any configuration, and the differences between sliding window sizes become more evident. Configurations with Inline graphic and Inline graphic consistently outperform those with Inline graphic and Inline graphic. For ResNet-18, all settings surpass 80% accuracy after 80 epochs, with Inline graphic again delivering the best performance. It is worth noting that the performance gap between Inline graphic and Inline graphic is most evident during the initial training phase, indicating that overly large sliding windows may inject high-variance information that negatively affects convergence speed and stability in complex tasks.

In summary, these results demonstrate that the sliding window size is a critical hyperparameter affecting model convergence and generalization, particularly for shallow architectures and complex datasets. Across all datasets and model configurations, a window size of Inline graphic consistently achieves superior or near-optimal results. Smaller window sizes (Inline graphic) may lead to insufficient contextual encoding, while larger windows (Inline graphic) risk incorporating irrelevant information and increasing overfitting potential. These findings offer empirical evidence to inform the selection of sliding window parameters in context-aware learning scenarios, especially those involving temporal or spatial dependencies.

To further evaluate the impact of sliding window size on model performance, Table 3 reports four key evaluation metrics—Accuracy, Precision, Recall, and F1-score—under different window lengths Inline graphic. As observed from Table 3, both LeNet-5 and ResNet-18 achieve strong performance across all metrics on the MNIST dataset, reflecting the simplicity of this task. LeNet-5 achieves optimal results at Inline graphic, with Accuracy (0.9721), Precision (0.9701), Recall (0.9714), and F1-score (0.9707) all reaching their highest values. As k increases to 5, performance slightly declines. A similar trend is observed in ResNet-18, which performs best at Inline graphic (Accuracy = 0.9805, F1 = 0.9774), with a slight decrease in all metrics occurring at Inline graphic. These results suggest that for MNIST, a moderate sliding window size like Inline graphic is most suitable, while overly large windows may introduce redundant spatial information, causing marginal performance degradation.

Table 3.

Comparison test of different sliding windows.

Dataset Model Window Accuracy Precision Recall F1
MNIST LeNet-5 2 0.9141 0.9133 0.9137 0.9135
3 0.9721 0.9701 0.9714 0.9707
4 0.9447 0.9436 0.9428 0.9431
5 0.9239 0.9233 0.9221 0.9226
ResNet18 2 0.9624 0.9608 0.9621 0.9614
3 0.9805 0.9763 0.9786 0.9774
4 0.9603 0.9610 0.9626 0.9618
5 0.9481 0.9516 0.9502 0.9509
Fashion MNIST LeNet-5 2 0.8030 0.8037 0.8040 0.8038
3 0.8515 0.8528 0.8537 0.8532
4 0.8317 0.8346 0.8320 0.8333
5 0.7918 0.7923 0.7935 0.7929
ResNet18 2 0.8624 0.8611 0.8615 0.8613
3 0.9129 0.9100 0.9123 0.9111
4 0.8902 0.8903 0.8914 0.8908
5 0.8316 0.8292 0.8315 0.8303
CIFAR-10 LeNet-5 2 0.5658 0.5634 0.5649 0.5641
3 0.6083 0.6139 0.6171 0.6155
4 0.5952 0.5943 0.5965 0.5954
5 0.5857 0.5867 0.5819 0.5843
ResNet18 2 0.8118 0.8105 0.8111 0.8108
3 0.8495 0.8515 0.8484 0.8499
4 0.8427 0.8414 0.8427 0.8420
5 0.8401 0.8381 0.8419 0.8400

On the more complex Fashion MNIST dataset, model sensitivity to window size becomes more pronounced. LeNet-5 achieves its best performance at Inline graphic, with the F1-score increasing from 0.8038 at Inline graphic to 0.8283 at Inline graphic, showing only a slight improvement at Inline graphic (0.8333) and a clear decline at Inline graphic (0.7923), particularly during early training stages. This suggests that an overly large sliding window may introduce noise or irrelevant features, impairing the generalization ability of shallow networks. ResNet-18 also achieves optimal results at Inline graphic on this dataset (Accuracy = 0.9129, F1 = 0.9111). Although performance at Inline graphic and Inline graphic is slightly lower, the model demonstrates greater stability overall while still benefiting from an appropriate window size.

CIFAR-10 is the most challenging of the three datasets due to its greater category diversity and structural complexity. On this task, LeNet-5 performs significantly worse than on the other datasets, with accuracy ranging from 0.5658 to 0.6083, and the best performance again observed at Inline graphic. In contrast, ResNet-18 exhibits substantially better results, achieving an accuracy of 0.8495 and F1-score of 0.8486 at Inline graphic. Although there is a slight performance drop at Inline graphic and Inline graphic, the overall metrics remain stable. This demonstrates that deep networks are more tolerant to changes in window size, yet Inline graphic still yields the best results under complex task conditions.

Overall, the results in Table 3 clearly indicate that sliding window size k is a crucial hyperparameter affecting model performance. Across all datasets and model architectures, Inline graphic consistently achieves optimal or near-optimal performance. Shallow models (e.g., LeNet-5) are more sensitive to changes in window size, while deeper models (e.g., ResNet-18) show greater robustness but still benefit from well-chosen window sizes. A too-small window (e.g., Inline graphic) may lead to insufficient contextual information, while a too-large window (e.g., Inline graphic) may introduce redundancy or noise, increasing the risk of overfitting and slowing convergence.

Performance comparison experiment between ALDP-FL and state-of-the-art methods

In this set of experiments, comparative tests were conducted on the ALDP-FL method against FedAvg, Fed-DPA, GFL-ALDPA, and LDP-Fed methods using different network models and datasets. Based on the previous set of experiments, the sliding window parameter k in the ALDP-FL method was set to 3.

As illustrated in Fig. 6, five methods—Fed-DPA, FedAvg, LDP-Fed, GFL-ALDPA, and ALDP-FL—are compared in terms of test accuracy across three benchmark datasets: MNIST, Fashion MNIST, and CIFAR-10. Two backbone models are employed: LeNet-5 (top row) and ResNet-18 (bottom row).On the MNIST dataset, all methods converge rapidly within the first 20 global epochs, and the accuracy stabilizes quickly. The baseline method FedAvg achieves the best performance in both LeNet-5 and ResNet-18 models, with accuracy consistently exceeding 95%. Among the privacy-preserving approaches, ALDP-FL achieves the best results, followed by GFL-ALDPA, while Fed-DPA remains the weakest. Notably, ALDP-FL significantly narrows the performance gap with FedAvg, especially when using the deeper ResNet-18 model. These results suggest that ALDP-FL achieves a favorable balance between privacy protection and model utility in relatively simple tasks.

Fig. 6.

Fig. 6

Comparison test of different methods.

The Fashion MNIST dataset poses greater complexity and class similarity. In this case, FedAvg still attains the highest accuracy; however, its advantage over privacy-enhanced methods becomes smaller. ALDP-FL outperforms all other privacy-preserving methods, further closing the gap with FedAvg. GFL-ALDPA ranks second, while Fed-DPA still performs the worst. With ResNet-18, the overall performance improves substantially, and ALDP-FL benefits more significantly from the deep model, yielding faster convergence and higher accuracy. These findings demonstrate that ALDP-FL is more robust to data complexity and scales well with deeper architectures.

CIFAR-10 represents the most challenging task in this study due to its high visual complexity and ambiguous semantic boundaries. The performance differences between methods become more pronounced. FedAvg still achieves the highest accuracy in both models. Among privacy-preserving approaches, ALDP-FL clearly performs best, achieving nearly 85% accuracy under ResNet-18. In contrast, Fed-DPA performs the worst, with accuracy failing to surpass 73% even under the deep model. These results indicate that ALDP-FL maintains strong performance in complex scenarios, making it more applicable to real-world deployments with privacy constraints.

Table 4 further presents the quantitative results of four key evaluation metrics—Accuracy, Precision, Recall, and F1 Score—across the five methods. The experimental results clearly demonstrate that ALDP-FL exhibits significant advantages across multiple performance indicators, with particularly strong performance in complex tasks and deep model architectures. To provide a more intuitive understanding of this advantage, we conducted a comparative analysis of the four core metrics and calculated the average performance improvement of ALDP-FL over other privacy-preserving methods.

Table 4.

Comparison test of different methods.

Dataset Model Approach Accuracy Precision Recall F1
MNIST LeNet-5 FedAvg 0.9961 0.9984 0.9957 0.9970
Fed-DPA 0.8735 0.8741 0.8745 0.8743
GFL-ALDPA 0.9218 0.9250 0.9226 0.9238
LDP-Fed 0.9150 0.9138 0.9144 0.9141
ALDP-FL 0.9725 0.9701 0.9750 0.9725
ResNet18 FedAvg 0.9996 0.9997 0.9986 0.9991
Fed-DPA 0.9213 0.9230 0.9237 0.9233
GFL-ALDPA 0.9512 0.9535 0.9536 0.9535
LDP-Fed 0.9531 0.9538 0.9547 0.9542
ALDP-FL 0.9823 0.9818 0.9781 0.9799
Fashion MNIST LeNet-5 FedAvg 0.9135 0.9146 0.9134 0.9140
Fed-DPA 0.7327 0.7318 0.7340 0.7329
GFL-ALDPA 0.8341 0.8343 0.8327 0.8335
LDP-Fed 0.8120 0.8115 0.8147 0.8131
ALDP-FL 0.8528 0.8519 0.8524 0.8521
ResNet18 FedAvg 0.9542 0.9537 0.9523 0.9530
Fed-DPA 0.8148 0.8099 0.8139 0.8119
GFL-ALDPA 0.8835 0.8832 0.8831 0.8831
LDP-Fed 0.8640 0.8649 0.8627 0.8638
ALDP-FL 0.9129 0.9123 0.9092 0.9107
CIFAR-10 LeNet-5 FedAvg 0.6931 0.6920 0.6946 0.6933
Fed-DPA 0.4841 0.4837 0.4829 0.4833
GFL-ALDPA 0.5738 0.5728 0.5725 0.5726
LDP-Fed 0.5449 0.5438 0.5441 0.5439
ALDP-FL 0.6132 0.6137 0.6171 0.6154
ResNet18 FedAvg 0.9146 0.9136 0.9141 0.9138
Fed-DPA 0.7241 0.7245 0.7233 0.7239
GFL-ALDPA 0.8040 0.8037 0.8028 0.8032
LDP-Fed 0.7533 0.7541 0.7534 0.7537
ALDP-FL 0.8526 0.8476 0.8518 0.8497

Statistical analysis shows that ALDP-FL achieves an average improvement of over 10% in all metrics: Accuracy increases by 10.57%, Precision by 10.64%, Recall by 10.52%, and F1 Score by 10.64%. Notably, in the most challenging dataset—CIFAR-10—ALDP-FL demonstrates a substantial F1 Score improvement of 9.60% over LDP-Fed when using the ResNet18 model. This highlights its effectiveness in handling fine-grained classification tasks and ambiguous object boundaries.

These findings suggest that ALDP-FL not only offers robust privacy protection but also effectively mitigates the performance degradation typically associated with conventional privacy-preserving techniques in image recognition tasks. In contrast, although Fed-DPA implements a basic differential privacy mechanism, it severely compromises model performance, with an average performance drop exceeding 17%. While GFL-ALDPA and LDP-Fed partially alleviate this issue, they still show noticeable deficits in terms of Accuracy and F1 Score.

In conclusion, ALDP-FL significantly outperforms existing privacy-preserving methods in terms of overall performance while maintaining strong privacy guarantees, making it a promising solution for privacy-sensitive federated learning scenarios.

Privacy protection capability testing experiment based on model inversion attacks

To evaluate the effectiveness of different privacy-preserving methods against reconstruction attacks, this study employs the improved Deep Leakage from Gradients (iDLG) algorithm to perform gradient leakage attacks on ResNet models trained on three benchmark datasets: MNIST, FashionMNIST, and CIFAR-10. The reconstructed images are visualized for comparative analysis. The experimental results are shown in Fig. 7.

Fig. 7.

Fig. 7

Visual results of the iDLG attack on the three datasets.

From the visualizations, the following observations can be made:In the MNIST dataset, the original images exhibit clearly recognizable digits with sharp edges. Under the FedAvg method, which does not incorporate any privacy protection, the attacker is able to almost perfectly reconstruct the original images, indicating that the model gradients contain a significant amount of recoverable sensitive information. In contrast, when privacy-preserving mechanisms such as Fed-DPA, GLF-ALDPA, LDP-Fed, and ALDP-FL are applied, the quality of the reconstructed images deteriorates significantly. The boundaries become blurred, and in many cases, the digits become unrecognizable. Particularly, ALDP-FL produces the most distorted reconstructions with no discernible digit structure, demonstrating its strong resistance to iDLG attacks.

In the FashionMNIST dataset, the images are more complex than MNIST, containing finer texture information. Similar to MNIST, the reconstructions under FedAvg still retain a relatively high degree of fidelity, with identifiable clothing contours. However, the other privacy methods introduce substantial interference to the reconstruction. Notably, GLF-ALDPA and ALDP-FL severely distort the reconstructed images across many samples, rendering them almost completely noisy and effectively suppressing information leakage.

For the CIFAR-10 dataset, which consists of color images with semantically complex content, a higher robustness against reconstruction is required. The original images clearly depict natural objects such as airplanes, horses, and ships. While FedAvg is still able to partially recover the global color structure of the images, the reconstructions are blurry and lack detail. The privacy-preserving methods, particularly ALDP-FL, demonstrate more effective protection in this scenario. The reconstructed images under ALDP-FL are nearly fully covered by color noise and are visually unrecognizable, exhibiting excellent leakage prevention capabilities.

Overall, the privacy-preserving methods show consistent trends across datasets: for grayscale datasets like MNIST and FashionMNIST, the methods strongly disrupt structural information; for color datasets like CIFAR-10, the defensive effect remains robust. Among these, ALDP-FL consistently achieves the best performance across all three datasets, maximally impairing the reconstructability under iDLG attacks and proving to be an effective approach for safeguarding image privacy.

To further quantify the protective effects of the methods, Table 5 presents four evaluation metrics of reconstructed image quality under iDLG attacks: Mean Squared Error (MSE), Peak Signal-to-Noise Ratio (PSNR), Mean Absolute Error (MAE), and Structural Similarity Index (SSIM). Using FedAvg as the baseline, we calculate the percentage change in performance for ALDP-FL across all metrics:On the MNIST dataset, ALDP-FL increases the MSE to 0.2226 (a 237.6% increase over FedAvg) and reduces the SSIM from 0.5368 to 0.0525 (a 90.% decrease), indicating that the reconstructed images have nearly lost all structural resemblance to the originals.On FashionMNIST, ALDP-FL increases the MSE from 0.0341 to 0.1852 (an increase of 443.7%) and decreases SSIM by 82.6%, further confirming its strong resistance to reconstruction in grayscale image scenarios.For the CIFAR-10 color image dataset, ALDP-FL achieves an MSE of 0.1836 (a 492.3% increase) and reduces SSIM to 0.0808 (an 83.5% decrease). Although MAE decreases, this does not significantly affect the overall structural disruption.These results demonstrate that ALDP-FL effectively increases the unrecognizability of reconstructed images across all three datasets. Specifically, its average improvement in MSE and SSIM reaches 391.2% and -85.4%, respectively, substantially outperforming all other compared methods.

Table 5.

Privacy evaluation metrics between original and reconstructed images under the iDLG attack.

Dataset Approach MSE PSNR MAE SSIM
MNIST FedAvg 0.0659 11.81 0.1609 0.5368
Fed-DPA 0.2025 6.93 0.2949 0.1891
GFL-ALDPA 0.2213 6.36 0.3234 0.0974
LDP-Fed 0.2088 6.80 0.2972 0.1090
ALDP-FL 0.2226 6.5 0.3004 0.0525
Fashion MNIST FedAvg 0.0341 14.67 0.1306 0.5085
Fed-DPA 0.1279 8.93 0.2624 0.2494
GFL-ALDPA 0.1317 8.8 0.2594 0.1468
LDP-Fed 0.1492 8.26 0.2745 0.1354
ALDP-FL 0.1852 7.32 0.2827 0.0882
CIFAR-10 FedAvg 0.0310 0.0753 15.08 0.4895
Fed-DPA 0.1003 0.2320 9.9 0.1870
GFL-ALDPA 0.1214 0.2716 9.16 0.1231
LDP-Fed 0.1062 0.2489 9.74 0.1403
ALDP-FL 0.1836 0.3352 7.36 0.0808

In summary, ALDP-FL exhibits the most robust and comprehensive defense performance against gradient-based reconstruction attacks. It significantly increase reconstruction error while destroying structural similarity, making it a highly effective and generally applicable solution for enhancing privacy protection in federated learning.

Conclusion

This paper addresses privacy leakage issues in federated learning by proposing ALDP-FL, a federated learning method based on adaptive local differential privacy. The algorithm is designed to mitigate privacy risks caused by malicious interception of sensitive data by central servers, real-time distributed attackers, or other adversaries. Additionally, a bounded noise mechanism is introduced to reduce the impact of noise on the accuracy of the global model. A series of experiments validate the effectiveness and practicality of ALDP-FL.

Future work will explore rational allocation of privacy budgets to enhance global model accuracy while preserving participant privacy, as well as extending differentially private federated learning methods to more real-world applications.

Author contributions

Lixin Cui:methodology,visualization,data curation, writing original draft preparation. Xu Wu:conceptualization, supervision, writing reviewing and editing.

Data availability

The datasets generated or analysed during the current study are available in this repository https://github.com/C-Kepler/Federated-Learning.git.

Declarations

Competing interests

The authors declare no competing interests.

Footnotes

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Thapa, C. & Camtepe, S. Precision health data: Requirements, challenges and existing techniques for data security and privacy. Comput. Biol. Med.129, 104130 (2021). [DOI] [PubMed] [Google Scholar]
  • 2.Janiesch, C., Zschech, P. & Heinrich, K. Machine learning and deep learning. Electron. Mark.31, 685–695 (2021). [Google Scholar]
  • 3.Sharifani, K. & Amini, M. Machine learning and deep learning: A review of methods and applications. World Inf. Technol. Eng. J.10, 3897–3904 (2023). [Google Scholar]
  • 4.Liu, Z. et al. Privacy-preserving aggregation in federated learning: A survey. IEEE Transactions on Big Data (2022).
  • 5.Banabilah, S., Aloqaily, M., Alsayed, E., Malik, N. & Jararweh, Y. Federated learning review: Fundamentals, enabling technologies, and future applications. Inf. Process. Manag.59, 103061 (2022). [Google Scholar]
  • 6.Li, L., Fan, Y., Tse, M. & Lin, K.-Y. A review of applications in federated learning. Comput. Ind. Eng.149, 106854 (2020). [Google Scholar]
  • 7.Liu, G., Wang, C., Ma, X. & Yang, Y. Keep your data locally: Federated-learning-based data privacy preservation in edge computing. IEEE Netw.35, 60–66 (2021). [Google Scholar]
  • 8.Jia, B. et al. Blockchain-enabled federated learning data protection aggregation scheme with differential privacy and homomorphic encryption in iiot. IEEE Trans. Ind. Inform.18, 4049–4058 (2021). [Google Scholar]
  • 9.Li, Q., He, B. & Song, D. Model-contrastive federated learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 10713–10722 (2021).
  • 10.Cinà, A. E. et al. Wild patterns reloaded: A survey of machine learning security against training data poisoning. ACM Comput. Surv.55, 1–39 (2023). [Google Scholar]
  • 11.Li, Q. et al. A survey on federated learning systems: Vision, hype and reality for data privacy and protection. IEEE Trans. Knowl. Data Eng.35, 3347–3366 (2021). [Google Scholar]
  • 12.Zhang, K., Song, X., Zhang, C. & Yu, S. Challenges and future directions of secure federated learning: a survey. Front. Comput. Sci.16, 1–8 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Yin, X., Zhu, Y. & Hu, J. A comprehensive survey of privacy-preserving federated learning: A taxonomy, review, and future directions. ACM Comput. Surv. (CSUR)54, 1–36 (2021). [Google Scholar]
  • 14.Gupta, P., Yadav, K., Gupta, B. B., Alazab, M. & Gadekallu, T. R. A novel data poisoning attack in federated learning based on inverted loss function. Comput. Secur.130, 103270 (2023). [Google Scholar]
  • 15.Gu, Y., Bai, Y. & Xu, S. Cs-mia: Membership inference attack based on prediction confidence series in federated learning. J. Inf. Secur. Appl.67, 103201 (2022). [Google Scholar]
  • 16.Zhu, L., Liu, Z. & Han, S. Deep leakage from gradients. Advances in neural information processing systems32 (2019).
  • 17.Hu, K. et al. An overview of implementing security and privacy in federated learning. Artif. Intell. Rev.57, 204 (2024). [Google Scholar]
  • 18.Gosselin, R., Vieu, L., Loukil, F. & Benoit, A. Privacy and security in federated learning: A survey. Appl. Sci.12, 9901 (2022). [Google Scholar]
  • 19.Adnan, M., Kalra, S., Cresswell, J. C., Taylor, G. W. & Tizhoosh, H. R. Federated learning and differential privacy for medical image analysis. Sci. Rep.12, 1953 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Ma, X., Zhu, J., Lin, Z., Chen, S. & Qin, Y. A state-of-the-art survey on solving non-iid data in federated learning. Fut. Gener. Comput. Syst.135, 244–258 (2022). [Google Scholar]
  • 21.Zhang, L., Zhu, T., Xiong, P., Zhou, W. & Yu, P. S. A robust game-theoretical federated learning framework with joint differential privacy. IEEE Trans. Knowl. Data Eng.35, 3333–3346 (2022). [Google Scholar]
  • 22.McMahan, H. B., Ramage, D., Talwar, K. & Zhang, L. Learning differentially private recurrent language models. arXiv preprint arXiv:1710.06963 (2017).
  • 23.Yang, J. et al. Gfl-aldpa: A gradient compression federated learning framework based on adaptive local differential privacy budget allocation. Multimed. Tools Appl.83, 26349–26368 (2024). [Google Scholar]
  • 24.Truex, S., Liu, L., Chow, K.-H., Gursoy, M. E. & Wei, W. Ldp-fed: Federated learning with local differential privacy. In Proceedings of the third ACM international workshop on edge systems, analytics and networking, 61–66 (2020).
  • 25.Yang, X., Huang, W. & Ye, M. Dynamic personalized federated learning with adaptive differential privacy. Adv. Neural Inf. Process. Syst.36, 72181–72192 (2023). [Google Scholar]
  • 26.Chen, C. et al. Trustworthy federated learning: privacy, security, and beyond. Knowl. Inf. Syst.67, 2321–2356 (2025). [Google Scholar]
  • 27.Bai, L. et al. Membership inference attacks and defenses in federated learning: A survey. ACM Comput. Surv.57(4), 1–35 (2024). [Google Scholar]
  • 28.Zhao, B., Mopuri, K. R. & Bilen, H. idlg: Improved deep leakage from gradients. arXiv preprint arXiv:2001.02610 (2020).
  • 29.Khraisat, A. et al. Securing federated learning: a defense strategy against targeted data poisoning attack. Discov Internet Things5, 16 (2025). [Google Scholar]
  • 30.Sharma, A. & Marchang, N. A review on client-server attacks and defenses in federated learning. Comput. Secur.140, 103801 (2024). [Google Scholar]
  • 31.Tripathy, R., Meshram, J. & Bera, P. Halffedlearn: A secure federated learning with local data partitioning and homomorphic encryption. Fut. Gener. Comput. Syst.171, 107858 (2025). [Google Scholar]
  • 32.Xie, Q. et al. Efficiency optimization techniques in privacy-preserving federated learning with homomorphic encryption: A brief survey. IEEE Internet Things J.11, 24569–24580 (2024). [Google Scholar]
  • 33.Aziz, R., Banerjee, S., Bouzefrane, S. & Le Vinh, T. Exploring homomorphic encryption and differential privacy techniques towards secure federated learning paradigm. Fut. Internet15, 310 (2023). [Google Scholar]
  • 34.Chang, Y., Zhang, K., Gong, J. & Qian, H. Privacy-preserving federated learning via functional encryption, revisited. IEEE Trans. Inf. Forensics Secur.18, 1855–1869 (2023). [Google Scholar]
  • 35.Park, J. & Lim, H. Privacy-preserving federated learning using homomorphic encryption. Appl. Sci.12, 734 (2022). [Google Scholar]
  • 36.Shan, F., Mao, S., Lu, Y. & Li, S. Differential privacy federated learning: A comprehensive review. Int. J. Adv. Comput. Sci. Appl.10.14569/IJACSA.2024.0150722 (2024). [Google Scholar]
  • 37.Xu, Z. et al. Federated learning of gboard language models with differential privacy. arXiv preprint arXiv:2305.18465 (2023).
  • 38.Xie, H., Zhang, Y., Zhongwen, Z. & Zhou, H. Privacy-preserving medical data collaborative modeling: A differential privacy enhanced federated learning framework. J. Knowl. Learn. Sci. Technol.3, 340–350 (2024). [Google Scholar]
  • 39.Fu, J. et al. Differentially private federated learning: A systematic review. arXiv preprint arXiv:2405.08299 (2024).
  • 40.Ling, J., Zheng, J. & Chen, J. Efficient federated learning privacy preservation method with heterogeneous differential privacy. Comput. Secur.139, 103715 (2024). [Google Scholar]
  • 41.Ren, X., Yang, S., Zhao, C., McCann, J. & Xu, Z. Belt and braces: When federated learning meets differential privacy. Commun. ACM67, 66–77 (2024). [Google Scholar]
  • 42.Banse, A., Kreischer, J. et al. Federated learning with differential privacy. arXiv preprint arXiv:2402.02230 (2024).
  • 43.Pakina, A. K. & Pujari, M. Differential privacy at the edge: A federated learning framework for gdpr-compliant tinyml deployments. IOSR J. Comput. Eng.26, 52–64 (2024). [Google Scholar]
  • 44.Wang, D. & Guan, S. Fedfr-adp: Adaptive differential privacy with feedback regulation for robust model performance in federated learning. Inf. Fusion116, 102796 (2025). [Google Scholar]
  • 45.Talaei, M. & Izadi, I. Adaptive differential privacy in federated learning: A priority-based approach. arXiv preprint arXiv:2401.02453 (2024).
  • 46.Beltrán, E. T. M. et al. Decentralized federated learning: Fundamentals, state of the art, frameworks, trends, and challenges. IEEE Commun. Surv. Tutor.25, 2983–3013 (2023). [Google Scholar]
  • 47.Reina, G. A. et al. Openfl: An open-source framework for federated learning. arXiv preprint arXiv:2105.06413 (2021).
  • 48.Cheng, K. et al. Secureboost: A lossless federated learning framework. IEEE Intell. Syst.36, 87–98 (2021). [Google Scholar]
  • 49.Rahman, K. J. et al. Challenges, applications and design aspects of federated learning: A survey. IEEE Access9, 124682–124700 (2021). [Google Scholar]
  • 50.Wang, N. et al. Collecting and analyzing multidimensional data with local differential privacy. In 2019 IEEE 35th International Conference on Data Engineering (ICDE), 638–649 (IEEE, 2019).
  • 51.Dong, J., Roth, A. & Su, W. J. Gaussian differential privacy. J. R. Stat. Soc.: Ser. B (Stat. Methodol.)84, 3–37 (2022). [Google Scholar]
  • 52.Zhang, J. et al. idp-fl: A fine-grained and privacy-aware federated learning framework for deep neural networks. Inf. Sci.679, 121035 (2024). [Google Scholar]
  • 53.Zheng, S. et al. Error-bounded correction of noisy labels. In International Conference on Machine Learning, 11447–11457 (PMLR, 2020).

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The datasets generated or analysed during the current study are available in this repository https://github.com/C-Kepler/Federated-Learning.git.


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES