Edge Computing brings computation and data storage closer to data sources, reducing latency and bandwidth usage.
Federated Learning, on the other hand, allows multiple devices to collaboratively train machine learning models without sharing raw data, preserving privacy.
Together, EC and FL facilitate decentralized intelligence, crucial for applications requiring immediate data processing and stringent privacy measures.
Given their complementary characteristics, edge computing provides an ideal environment for deploying federated learning. This natural synergy has made Edge Federated Learning an increasingly attractive solution for both academic research and industrial applications.
In this context, it is essential to explore the fundamentals and unique benefits of both technologies to fully understand their combined potential.
The rise of real-time technologies such as virtual reality (VR), augmented reality (AR), and autonomous vehicles has prompted researchers and industry leaders to explore new data processing architectures.
Traditional cloud computing models are not well-suited for applications that require ultra-low latency. This limitation has led to the emergence of Edge Computing (EC), a paradigm focused on processing and transmitting data closer to the source, near end-user applications, rather than relying on centralized servers.
In this framework, edge nodes (also known as edge clients or edge devices) are typically user-operated devices with limited resources, located geographically close to edge servers that offer greater computational power and bandwidth. When additional resources are needed, edge servers can offload tasks to cloud servers.
There are several compelling reasons why industry leaders are increasingly shifting away from traditional cloud-based models in favor of edge computing platforms.
Federated Learning (FL) has gained substantial attention as a collaborative machine learning approach where multiple devices, each with its own private dataset, train a shared global model without exchanging raw data.
FL not only allows distributed utilization of computing resources but also ensures user privacy is maintained throughout the learning process.
Given their complementary characteristics, edge computing provides an ideal environment for deploying federated learning. This natural synergy has made Edge Federated Learning an increasingly attractive solution for both academic research and industrial applications.
Federated Learning (FL) generally follows a structured, iterative process involving collaboration between a central server and multiple distributed clients (e.g., mobile devices, organizations, or edge nodes). Here are the main stages of federated learning:
This cycle continues for multiple rounds until the model reaches a satisfactory level of performance or convergence.
Edge Federated Learning (EFL) addresses the data silo problem by effectively utilizing the vast data generated on end-user devices, all while preserving user privacy.
This approach significantly enhances the efficiency of model training within edge computing environments, making it highly suitable for scenarios where both privacy protection and resource optimization are essential.
However, designing an EFL system involves multiple complexities. Developers must carefully account for factors such as varying APIs, data flow models, network configurations, and device capabilities.
Given these challenges, there is a growing need for the research community to focus on creating tools and frameworks that simplify the development and deployment of edge federated learning systems, enabling more efficient and accessible implementation.
The architecture is composed of three fundamental layers:
which together form the backbone of an edge computing-based federated learning system.
Unlike traditional federated learning, Edge Federated Learning (EFL) introduces an intermediate layer for distributed training. In EFL:
Just like traditional federated learning, edge federated learning can also be divided into two main categories based on the scale of participation: cross-silo and cross-device edge federated learning.
Cross-device federated learning can be significantly enhanced when integrated with a cloud-edge collaborative framework.
Unlike traditional cross-device FL, this architecture leverages edge servers to offer additional protection for local network traffic. If an edge server is compromised, it can be removed from the training process without affecting the rest of the system, thus isolating attacks and preserving the integrity of other devices’ training processes.
Beyond improved security, edge servers can also safely handle computational tasks offloaded from client devices, providing low-latency processing capabilities. This is especially beneficial for mobile devices, which may have limited processing power but maintain strong communication links with nearby edge servers, reducing their computational load during model training.
In cross-device edge federated learning, the number of participating training nodes can reach into the millions, with each node typically possessing only a small amount of data and limited computational capability. These nodes are generally portable devices or sensors, such as smartphones, smartwatches, or IoT-enabled equipment.
In the cloud-edge collaborative model, cross-silo FL gains additional flexibility and efficiency. In this context, each silo’s dataset is better treated as an individual learning task rather than just part of a fragmented global dataset.
One of the key challenges in cross-silo FL is the Non-IID (non-identically distributed) nature of data across different clients. Traditionally, meta-learning and transfer learning techniques are used within centralized cloud environments to address this. However, cloud-edge collaboration introduces an alternative: Hierarchical Federated Learning through clustering.
This method involves grouping clients based on similarities in their data distribution, with each group managed by a dedicated edge server. Studies have validated that this cluster-based approach is effective in mitigating Non-IID challenges, leading to better model performance and more efficient federated training.
In cross-silo edge federated learning, the number of participating nodes is relatively small, but each node, typically an edge server, must be equipped with ample computational resources to handle large-scale datasets. For instance, major e-commerce platforms may use this setup to recommend products by training on tens of millions of shopping records distributed across geographically separated data centers.
Security and privacy are two of the most critical challenges in implementing federated learning (FL) within edge computing environments.
Due to the inherently heterogeneous and decentralized nature of both technologies, predicting the behavior of participating nodes becomes difficult. During training, some edge nodes may act maliciously and attempt to compromise the learning process.
Moreover, curious or semi-honest nodes and servers may try to infer private information from the uploaded model updates, potentially compromising user privacy.
Security issues in Edge Federated Learning (EFL) stem largely from the lack of trust among the various components of the system.
Edge nodes and servers often come from different sources, and not all of them can be trusted. While federated learning enhances user privacy by keeping data local, this same feature makes it more vulnerable to malicious behaviors.
These are especially problematic in distributed learning. In the edge computing context, their impact is amplified due to:
Common Byzantine attack methods include:
Another major security threat in EFL is poisoning attacks, which aim to disrupt or compromise the model by manipulating inputs or local models. They are classified into:
Despite the different attack vectors, both Byzantine and poisoning attacks highlight the need for the central server to distinguish between honest and malicious updates during model aggregation.
To defend against such attacks, researchers have proposed Byzantine-resilient algorithms, which aim to filter or correct malicious updates. These generally fall into three categories:
Together, these defenses aim to enhance the security and robustness of federated learning in edge environments, enabling privacy-preserving intelligence even in untrusted, heterogeneous networks.
Although federated learning is designed to protect individual nodes by keeping their training data local and avoiding direct data transmission, privacy breaches can still occur during the sharing of model information, such as model weights, between nodes and servers.
In particular, curious or semi-honest participants, whether they are edge nodes or central servers, may attempt to infer sensitive data through the training process. In edge federated learning, the data exchanged typically includes locally computed model updates and aggregated weights from the server. As a result, attackers can potentially extract private information by analyzing these weight updates, either directly from the data uploaded by nodes (in the case of a curious server) or from the aggregated model parameters (in the case of a curious node).
This risk essentially boils down to reconstructing private training data from model updates. Two common types of privacy attacks in federated learning include:
To address these risks, privacy-preserving techniques in edge federated learning typically fall into two main categories:
Many advanced privacy-preserving strategies today fall into these two main categories. However, some solutions combine multiple techniques from both categories.
This article provided an overview of edge federated learning, emphasizing its role in enabling decentralized, privacy-preserving machine learning across resource-constrained edge environments. Information presented here integrates insights from recent academic research to offer accurate, current perspectives.
We examined the architectural layers, classifications (cross-device and cross-silo), and the integration of federated learning into edge computing systems.
We also explored key challenges related to communication efficiency, system scalability, and privacy threats, alongside mitigation techniques such as differential privacy, secure multi-party computation, and Byzantine-resilient aggregation.
As research continues to advance, edge federated learning stands out as a promising paradigm for intelligent, secure, and efficient AI at the edge.