Description
Federated Learning (FL) is a decentralized machine learning framework standardized within 3GPP to enable artificial intelligence (AI) and machine learning (ML) model training across the mobile network ecosystem while addressing data privacy and transmission constraints. In a typical 3GPP FL architecture, a central server, known as the Federated Learning Server (FL Server), coordinates the training process. This server initializes a global ML model (e.g., for radio resource management, mobility optimization, or service quality prediction) and distributes this model to participating FL Clients. These clients are typically User Equipments (UEs), but can also be network functions like base stations (gNBs) or edge computing nodes.
The core operational process involves multiple rounds of collaboration. In each round, the FL Server selects a set of clients and sends them the current global model. Each selected client then performs local training on its own private dataset, which never leaves the device. This local training computes an update to the model, typically in the form of model weights or gradients. Only this compact model update, not the raw data, is sent back to the FL Server. The server then aggregates all received updates (using algorithms like Federated Averaging) to produce an improved global model. This cycle repeats, progressively refining the global model based on the collective knowledge of all participating clients' data distributions.
Key components in the 3GPP FL system include the FL Client (the entity performing local training), the FL Server (orchestrating the process), and the FL Management System which handles client selection, resource provisioning, and lifecycle management. 3GPP specifications define the enabling protocols and interfaces, such as service-based interfaces for FL management and data transfer. The role of FL in the network is transformative, allowing for the creation of intelligent network and service functions that learn from real-world, distributed data generated at the edge—such as channel conditions, mobility patterns, or application usage—without compromising user privacy or overwhelming the transport network with massive data transfers. It turns the entire network of devices into a collective, privacy-aware AI training engine.
Purpose & Motivation
Federated Learning was introduced into 3GPP standards to solve two major problems inherent in centralizing data for network AI: data privacy/sovereignty and massive data transmission overhead. Traditional cloud-based ML requires aggregating vast amounts of raw user and network data in a central data center, raising significant privacy concerns, regulatory hurdles (like GDPR), and security risks from data breaches. Furthermore, transmitting all raw data from billions of UEs and network nodes to a central cloud is prohibitively expensive in terms of network bandwidth and latency.
The motivation for its creation stems from the industry's push towards embedded intelligence (Network Data Analytics Function - NWDAF, AI/ML in 5G-Advanced and 6G) and the need to leverage the exponentially growing data at the network edge. Previous approaches either ignored this distributed data or attempted complex and often non-compliant data anonymization and aggregation techniques. FL provides a fundamental architectural shift. It addresses these limitations by moving the computation to the data, rather than moving the data to the computation. This allows 3GPP networks to build accurate, generalized AI models for optimization and automation—such as predicting cell load, managing handovers, or detecting anomalies—by learning directly from user experiences and network conditions on devices and base stations, all while keeping sensitive information locally stored. It enables privacy-preserving collaboration on a scale necessary for future autonomous networks.
Key Features
- Decentralized model training across UEs and network nodes without raw data exchange
- Privacy-preserving by design, aligning with data protection regulations (GDPR)
- Reduces core network traffic by transmitting only compact model updates
- Supports heterogeneous data distributions across clients (non-IID data)
- Includes client selection, resource negotiation, and secure aggregation mechanisms
- Enables real-time network and service optimization using edge-generated data
Evolution Across Releases
Federated Learning was not introduced in Rel-8. The term 'FL' in early releases like Rel-8 through Rel-14 within the provided specs (e.g., 21.905, 22.261) refers to other concepts, such as 'Frequency Layer' or other legacy terms. The initial architecture and capabilities for Federated Learning as an AI/ML framework were not defined until later releases.
Federated Learning was formally introduced and studied as a key enabler for AI/ML in 5G-Advanced within the context of the 5G System. Work began on architectural frameworks, requirements, and use cases for applying FL to network data analytics, service optimization, and radio resource management, laying the groundwork for normative specifications.
Standardization of Federated Learning advanced significantly in 5G-Advanced. Normative work defined the FL management architecture, including the roles of FL Client, FL Server, and FL Management System. Protocols for client selection, model distribution, update aggregation, and resource provisioning were developed, integrating FL into the 3GPP service-based architecture.
Enhancements focused on FL efficiency and broader application. This included support for more sophisticated aggregation algorithms, improved handling of straggler clients, enhanced security for the aggregation process, and exploration of FL for new use cases like joint communication and sensing (JCAS) and further network automation.
Defining Specifications
| Specification | Title |
|---|---|
| TS 21.905 | 3GPP TS 21.905 |
| TS 22.261 | 3GPP TS 22.261 |
| TS 22.874 | 3GPP TS 22.874 |
| TS 23.288 | 3GPP TS 23.288 |
| TS 23.482 | 3GPP TS 23.482 |
| TS 23.501 | 3GPP TS 23.501 |
| TS 23.700 | 3GPP TS 23.700 |
| TS 24.560 | 3GPP TS 24.560 |
| TS 26.927 | 3GPP TS 26.927 |
| TS 28.105 | 3GPP TS 28.105 |
| TS 28.858 | 3GPP TS 28.858 |
| TS 29.482 | 3GPP TS 29.482 |
| TS 29.520 | 3GPP TS 29.520 |
| TS 29.552 | 3GPP TS 29.552 |
| TS 33.501 | 3GPP TR 33.501 |
| TS 33.700 | 3GPP TR 33.700 |
| TS 37.814 | 3GPP TR 37.814 |
| TS 45.903 | 3GPP TR 45.903 |