TLDR: FedEL is a new federated learning framework designed to improve training efficiency and accuracy on devices with diverse hardware capabilities. It uses a novel window-based training process that dynamically selects and trains parts of a machine learning model, ensuring all components are updated over time, even on slower devices. Additionally, FedEL adjusts tensor importance by considering both local and global model insights, mitigating biases from varied data distributions. This approach significantly speeds up training (up to 3.87x faster than traditional methods) while maintaining or improving model accuracy, and also reduces memory and energy consumption.
Federated learning (FL) represents a groundbreaking approach in machine learning, allowing numerous distributed devices, such as mobile phones and IoT systems, to collaboratively train a shared global model. A key advantage of FL is its privacy-preserving nature, as sensitive user data remains on local devices and only model parameters or updates are shared with a central server. This distributed paradigm is crucial for applications where data privacy and security are paramount.
The Challenge of Device Heterogeneity
Despite its promise, federated learning faces a significant hurdle: the vast heterogeneity in hardware capabilities among participating devices. In real-world scenarios, clients can range from powerful servers to resource-constrained edge devices. This disparity often leads to a problem known as “stragglers” – slower clients with limited computational resources that prolong the overall training process. The central server must wait for all clients, including the slowest ones, to complete their local training before aggregating updates, leading to substantial delays and hindering scalability.
Existing solutions attempt to address this, including client selection (choosing a subset of faster clients), asynchronous FL (allowing clients to update independently), and partial training (adjusting model architecture to fit resources). However, these methods often come with trade-offs, such as reduced model accuracy due to underrepresentation of stragglers, stale updates, or compromised model performance from inconsistent training contributions.
Introducing FedEL: A Novel Approach
To overcome these limitations, researchers have proposed FedEL, a federated elastic learning framework designed to enhance training efficiency while rigorously maintaining or even improving model accuracy. FedEL introduces two core innovations: a novel window-based training process and a tensor importance adjustment module.
Window-Based Training: A Dynamic View of the Model
FedEL’s window-based training addresses the issue of slower clients only training the latter parts of a deep neural network (DNN) model, which can degrade overall accuracy, especially with diverse data. Instead, FedEL divides the DNN model into multiple logical blocks. During each FL round, a “window” slides across these blocks, dynamically selecting a portion of the model for training within a coordinated runtime budget. This ensures that all parts of the DNN model, including the crucial front-end feature extraction layers, are progressively and equally trained across all clients, including stragglers.
To make this possible, FedEL first uses an offline tensor timing profiler to measure the training time for each tensor, which is then aggregated into block-level training times. The window’s movement is controlled by two boundaries: a “front edge” that moves forward to include deeper blocks based on cumulative training time, and an “end edge” that shrinks the window by excluding blocks that no longer require training or have no important tensors selected. A clever addition is a lightweight “early exit” output layer attached to the last layer of the current window, allowing each window to be trained independently without relying on the full model’s output layer. This iterative and cyclical process guarantees consistent training times for all clients while ensuring the entire model is thoroughly trained.
Tensor Importance Adjustment: Harmonizing Local and Global Insights
Another challenge in FL is “local model drift,” where models trained on diverse local datasets can diverge significantly from the global model. This is exacerbated when only “important” tensors are selected for training, as their importance can be biased by local data distributions. FedEL tackles this with a tensor importance adjustment module.
At the beginning of each FL round, clients estimate the global model’s tensor importance using the global models from the current and previous rounds, along with the learning rate. This “global tensor importance” is then used to adjust the “local tensor importance” calculated by ElasticTrainer (a component FedEL builds upon). By balancing local and global importance, FedEL ensures that tensor selection considers both individual client data characteristics and the overall global model priorities, thereby mitigating biases and improving the model’s overall training accuracy.
Impressive Performance and Efficiency Gains
The effectiveness of FedEL was rigorously evaluated through experiments on both a hardware testbed (using NVIDIA Jetson devices) and large-scale software simulations. It was tested across various DNN models and four real-world FL datasets for tasks like image classification (CIFAR10, Tiny ImageNet), voice command recognition (Google command speech), and next-word prediction (Reddit dataset). The datasets were partitioned to simulate realistic non-iid (non-independent and identically distributed) data scenarios.
The results are compelling: FedEL consistently outperformed baseline methods in terms of time-to-accuracy, achieving up to a 3.87 times improvement compared to FedAvg, while maintaining or even exceeding the final test accuracy. For instance, on image classification with 100 devices, FedEL achieved 3.59x speedup over FedAvg. Beyond speed and accuracy, FedEL also demonstrated significant reductions in memory overhead (up to 32.7% compared to FedAvg) and energy consumption (an average reduction of 49.59% in total energy usage), making it highly suitable for resource-constrained edge devices.
Ablation studies further confirmed the critical role of each component within FedEL’s design, including the balancing parameter for tensor importance adjustment and the dynamic sliding window mechanism. FedEL’s adaptive nature allows it to select important tensors dynamically, even skipping less important ones to maintain desired training speed while preserving model effectiveness.
Also Read:
- Boosting DNN Efficiency: The Power of Joint Memory and Computing Frequency Adjustments
- Securing Federated Learning in 5G and Edge Networks with Hybrid Reputation Tracking
Conclusion
FedEL represents a significant advancement in federated learning, offering a robust solution to the challenges posed by heterogeneous devices. By intelligently combining window-based training with a sophisticated tensor importance adjustment, it ensures efficient, accurate, and balanced training across diverse client environments. This framework paves the way for more scalable and practical deployments of federated learning in real-world applications. You can read the full research paper here.


