FL

Federated Learning

Other
Introduced in Rel-8
Federated Learning is a distributed machine learning paradigm where a global model is trained collaboratively across multiple decentralized devices or network nodes (like UEs or base stations) holding local data samples. It enables AI model training without exchanging raw data, preserving privacy and reducing network load.

Description

Federated Learning (FL) is a decentralized machine learning framework standardized within 3GPP to enable artificial intelligence (AI) and machine learning (ML) model training across the mobile network ecosystem while addressing data privacy and transmission constraints. In a typical 3GPP FL architecture, a central server, known as the Federated Learning Server (FL Server), coordinates the training process. This server initializes a global ML model (e.g., for radio resource management, mobility optimization, or service quality prediction) and distributes this model to participating FL Clients. These clients are typically User Equipments (UEs), but can also be network functions like base stations (gNBs) or edge computing nodes.

The core operational process involves multiple rounds of collaboration. In each round, the FL Server selects a set of clients and sends them the current global model. Each selected client then performs local training on its own private dataset, which never leaves the device. This local training computes an update to the model, typically in the form of model weights or gradients. Only this compact model update, not the raw data, is sent back to the FL Server. The server then aggregates all received updates (using algorithms like Federated Averaging) to produce an improved global model. This cycle repeats, progressively refining the global model based on the collective knowledge of all participating clients' data distributions.

Key components in the 3GPP FL system include the FL Client (the entity performing local training), the FL Server (orchestrating the process), and the FL Management System which handles client selection, resource provisioning, and lifecycle management. 3GPP specifications define the enabling protocols and interfaces, such as service-based interfaces for FL management and data transfer. The role of FL in the network is transformative, allowing for the creation of intelligent network and service functions that learn from real-world, distributed data generated at the edge—such as channel conditions, mobility patterns, or application usage—without compromising user privacy or overwhelming the transport network with massive data transfers. It turns the entire network of devices into a collective, privacy-aware AI training engine.

Purpose & Motivation

Federated Learning was introduced into 3GPP standards to solve two major problems inherent in centralizing data for network AI: data privacy/sovereignty and massive data transmission overhead. Traditional cloud-based ML requires aggregating vast amounts of raw user and network data in a central data center, raising significant privacy concerns, regulatory hurdles (like GDPR), and security risks from data breaches. Furthermore, transmitting all raw data from billions of UEs and network nodes to a central cloud is prohibitively expensive in terms of network bandwidth and latency.

The motivation for its creation stems from the industry's push towards embedded intelligence (Network Data Analytics Function - NWDAF, AI/ML in 5G-Advanced and 6G) and the need to leverage the exponentially growing data at the network edge. Previous approaches either ignored this distributed data or attempted complex and often non-compliant data anonymization and aggregation techniques. FL provides a fundamental architectural shift. It addresses these limitations by moving the computation to the data, rather than moving the data to the computation. This allows 3GPP networks to build accurate, generalized AI models for optimization and automation—such as predicting cell load, managing handovers, or detecting anomalies—by learning directly from user experiences and network conditions on devices and base stations, all while keeping sensitive information locally stored. It enables privacy-preserving collaboration on a scale necessary for future autonomous networks.

Key Features

  • Decentralized model training across UEs and network nodes without raw data exchange
  • Privacy-preserving by design, aligning with data protection regulations (GDPR)
  • Reduces core network traffic by transmitting only compact model updates
  • Supports heterogeneous data distributions across clients (non-IID data)
  • Includes client selection, resource negotiation, and secure aggregation mechanisms
  • Enables real-time network and service optimization using edge-generated data

Evolution Across Releases

Rel-8 Initial

Federated Learning was not introduced in Rel-8. The term 'FL' in early releases like Rel-8 through Rel-14 within the provided specs (e.g., 21.905, 22.261) refers to other concepts, such as 'Frequency Layer' or other legacy terms. The initial architecture and capabilities for Federated Learning as an AI/ML framework were not defined until later releases.

Federated Learning was formally introduced and studied as a key enabler for AI/ML in 5G-Advanced within the context of the 5G System. Work began on architectural frameworks, requirements, and use cases for applying FL to network data analytics, service optimization, and radio resource management, laying the groundwork for normative specifications.

Standardization of Federated Learning advanced significantly in 5G-Advanced. Normative work defined the FL management architecture, including the roles of FL Client, FL Server, and FL Management System. Protocols for client selection, model distribution, update aggregation, and resource provisioning were developed, integrating FL into the 3GPP service-based architecture.

Enhancements focused on FL efficiency and broader application. This included support for more sophisticated aggregation algorithms, improved handling of straggler clients, enhanced security for the aggregation process, and exploration of FL for new use cases like joint communication and sensing (JCAS) and further network automation.

Defining Specifications

SpecificationTitle
TS 21.905 3GPP TS 21.905
TS 22.261 3GPP TS 22.261
TS 22.874 3GPP TS 22.874
TS 23.288 3GPP TS 23.288
TS 23.482 3GPP TS 23.482
TS 23.501 3GPP TS 23.501
TS 23.700 3GPP TS 23.700
TS 24.560 3GPP TS 24.560
TS 26.927 3GPP TS 26.927
TS 28.105 3GPP TS 28.105
TS 28.858 3GPP TS 28.858
TS 29.482 3GPP TS 29.482
TS 29.520 3GPP TS 29.520
TS 29.552 3GPP TS 29.552
TS 33.501 3GPP TR 33.501
TS 33.700 3GPP TR 33.700
TS 37.814 3GPP TR 37.814
TS 45.903 3GPP TR 45.903