FOV

Field of View

Services
Introduced in Rel-14
Field of View (FOV) defines the extent of the observable environment that can be captured by a camera or sensor, measured as an angular width. In 3GPP, it is a critical parameter for immersive media services like augmented reality (AR), virtual reality (VR), and multi-view video, impacting bandwidth and quality requirements.

Description

In 3GPP specifications, Field of View (FOV) is a key parameter within media delivery and immersive service frameworks. It is an angular measurement, typically in degrees horizontally and vertically, that describes the portion of a full spherical or panoramic scene that is rendered or transmitted to a user's device at a given moment. Technically, it defines the frustum of vision for a viewpoint. For services like 360-degree video, the entire scene is captured (spherical mapping), but only the FOV corresponding to the user's head orientation is delivered to reduce bandwidth.

The architecture for FOV-aware media delivery involves components in the application server, media encoder, and the client. The media server typically stores or generates a high-resolution omnidirectional media stream. Based on viewport information (head tracking data) sent from the client, the server or a network-based processing function dynamically adapts the streaming. This can involve extracting and encoding only the tiles within the user's current FOV at high quality, while other tiles are sent at lower quality or not at all, a technique known as viewport-adaptive streaming.

Its role is central to enabling efficient delivery of bandwidth-intensive immersive media over mobile networks. By tracking the FOV and predicting future viewports, the system optimizes resource utilization on both the radio interface and the core network. This ensures high visual quality where the user is looking while minimizing latency and stutter when the view changes. Specifications such as those for MPEG Media Transport and Dynamic Adaptive Streaming over HTTP (DASH) have been extended to support FOV metadata and tile-based representations to enable this functionality.

Purpose & Motivation

The standardization of FOV-related parameters in 3GPP was driven by the emergence of commercial VR and AR services, which require massive amounts of video data for a fully immersive experience. Transmitting entire 360-degree scenes at ultra-high resolution would overwhelm network capacity and device processing power. The purpose of defining FOV mechanisms was to solve this scalability problem.

It addresses the limitation of traditional video streaming, which sends a single rectangular frame for all users. For immersive media, this is highly inefficient as most of the transmitted pixels fall outside the user's immediate view. By incorporating FOV into the service and transport layers, 3GPP standards enable a fundamental shift to viewport-dependent delivery. This makes high-quality, low-latency VR/AR experiences technically and economically feasible over 5G networks, supporting new use cases in gaming, remote collaboration, live events, and training.

Key Features

  • Angular measurement defining visible scene portion
  • Key enabler for viewport-adaptive streaming
  • Integrated with head orientation tracking data
  • Reduces bandwidth consumption for immersive media
  • Supports tile-based media encoding and delivery
  • Defined in media format and transport specs (e.g., MPEG, DASH)

Evolution Across Releases

Rel-14 Initial

Initial inclusion of FOV as a service parameter for immersive media, particularly for VR. Specifications began to define requirements for media formats and streaming capable of supporting viewport-dependent delivery to enable 360-degree video services over mobile networks.

Defining Specifications

SpecificationTitle
TS 22.156 3GPP TS 22.156
TS 22.856 3GPP TS 22.856
TS 26.114 3GPP TS 26.114
TS 26.118 3GPP TS 26.118
TS 26.862 3GPP TS 26.862
TS 26.918 3GPP TS 26.918
TS 26.929 3GPP TS 26.929