spot_img
HomeResearch & DevelopmentNew Framework Addresses Data Forgetting in Federated Learning with...

New Framework Addresses Data Forgetting in Federated Learning with Information Theory

TLDR: Researchers introduce a novel Federated Unlearning (FU) framework that tackles data erasure as a parameter estimation problem. By using second-order Hessian information, the method identifies and selectively resets only the model parameters most sensitive to the data to be forgotten, followed by minimal federated retraining. This approach demonstrates strong privacy (MIA success near random, categorical knowledge erased) and high performance (Normalized Accuracy against re-trained benchmarks of ≈0.9), while also effectively neutralizing backdoor attacks. Although efficiency can be a limitation for models with few training epochs, the framework scales better for more complex scenarios, offering a practical solution for the ‘right to be forgotten’ in federated learning.

In the rapidly evolving landscape of artificial intelligence, the ability to forget specific data from trained models has become a critical challenge, especially with the rise of stringent privacy regulations like GDPR and CCPA. These regulations grant users the “right to be forgotten,” demanding that their data be completely removed from any system, including the sophisticated deep learning models that power many modern applications.

This challenge is particularly amplified in Federated Learning (FL), a distributed approach where models are trained collaboratively across many client devices without centralizing raw data. While FL offers inherent privacy benefits by keeping data local, retroactively removing a user’s data contribution – a process known as Federated Unlearning (FU) – presents significant hurdles. The traditional “gold standard” of unlearning, which involves completely retraining the model from scratch without the forgotten data, is often impractical due to its high computational cost, communication overhead, and resource inefficiency, especially in large-scale FL deployments.

A Novel Approach to Federated Unlearning

A new research paper, titled Tackling Federated Unlearning as a Parameter Estimation Problem, introduces an innovative framework that redefines federated unlearning. Authored by Antonio Balordi, Lorenzo Manini, Fabio Stella, and Alessio Merlo, this work conceptualizes the problem of information leakage as a parameter estimation challenge, drawing on principles from information theory. Instead of full retraining, their method focuses on identifying and selectively resetting only the model parameters most sensitive to the data intended for removal.

The core of their approach lies in utilizing “second-order Hessian information.” In simpler terms, this involves analyzing how much each parameter in the neural network contributes to memorizing specific data. By computing a “Target Information Score” (TIS) for each parameter, the framework can pinpoint the most influential parameters related to the data that needs to be forgotten.

How the Unlearning Process Works

The proposed algorithm follows a two-stage process:

  1. Identification and Reset: After a model has been trained, clients collaboratively compute specific second-order statistics (Hessian diagonals) related to both the data to be forgotten and the data to be retained. This information is then aggregated by the central server, which calculates the TIS for each parameter. Based on a predefined “removal percentage,” the parameters with the highest TIS are identified and reset to their initial, pre-training random values. Crucially, this step does not require the server to access any raw client data.
  2. Minimal Retraining: To restore the model’s overall performance on the remaining data, a brief fine-tuning phase is performed. This phase uses a custom module called TRIM (Targeted Retraining via Index Masking). During this minimal federated retraining, which typically lasts for just a single epoch, only the reset parameters are updated using data exclusively from the remaining dataset. The non-reset parameters are kept frozen, ensuring that the influence of the forgotten data is minimized.

This model-agnostic approach is highly versatile, supporting various unlearning objectives such as removing specific data samples, an entire client’s dataset, or all data belonging to a particular class. It also works across diverse network architectures.

Key Advantages and Performance

The researchers evaluated their framework on benchmark datasets like MNIST, FashionMNIST, and CIFAR10, under different client data distribution strategies (random and preferential class settings). They also tested its resilience against a targeted backdoor attack scenario.

The results are compelling:

  • Strong Privacy: The unlearned models demonstrated robust privacy, with Membership Inference Attack (MIA) success rates approaching random guessing. This indicates that the models effectively forgot the target data, making it difficult for an attacker to determine if specific data was part of the training set. For categorical unlearning, the model’s ability to recognize the forgotten class dropped to near zero, signifying almost perfect forgetting.
  • High Performance: Despite forgetting, the unlearned models maintained high performance on general test data. The Normalized Test Accuracy (NTA) values were consistently close to 0.9 or higher, indicating that the unlearned model’s performance closely approximated that of a model ideally retrained from scratch.
  • Backdoor Neutralization: In a simulated backdoor attack, where malicious triggers were inserted into training data, the framework effectively neutralized the attack. The Attack Success Rate (ASR) was reduced to near-zero, and the model’s ability to correctly classify original images was fully restored, demonstrating its capability to remove malicious behavior and restore model integrity.

Also Read:

Efficiency and Future Directions

While the framework offers significant privacy and performance benefits, the researchers acknowledge a limitation regarding efficiency for models trained with a small number of epochs. The Recovery Time Ratio (RTR) indicates that for such scenarios, the unlearning method can be computationally more intensive than full retraining. However, the method scales better for larger models and tasks requiring more training epochs, where the cost of full retraining grows linearly with epochs, while their unlearning cost grows much more slowly.

Future work aims to address this efficiency aspect through optimized Hessian computation methods and further experiments on larger models. The researchers also plan to explore richer second-order information beyond Hessian diagonals and investigate adaptive hyperparameter tuning for the unlearning percentages to optimize the privacy-performance trade-offs.

Overall, this research presents a practical and principled solution for data forgetting in federated learning, offering a scalable and effective mechanism to meet the growing demands of data privacy regulations in distributed AI systems.

Ananya Rao
Ananya Raohttps://blogs.edgentiq.com
Ananya Rao is a tech journalist with a passion for dissecting the fast-moving world of Generative AI. With a background in computer science and a sharp editorial eye, she connects the dots between policy, innovation, and business. Ananya excels in real-time reporting and specializes in uncovering how startups and enterprises in India are navigating the GenAI boom. She brings urgency and clarity to every breaking news piece she writes. You can reach her out at: [email protected]

- Advertisement -

spot_img

Gen AI News and Updates

spot_img

- Advertisement -