top of page

Cloud & NFV(Network Function Virtualization) for 5G Networks

  • Writer: Venkateshu
    Venkateshu
  • a few seconds ago
  • 9 min read

1.Introduction

Cloud and Network Function Virtualization (NFV) for 5G networks represent a fundamental shift in how telecommunication networks are designed and managed. NFV decouples network functions from specialized hardware, allowing them to run as software on standard cloud infrastructure. This virtualization enables 5G networks to be more flexible, scalable, and cost-efficient by leveraging cloud-native principles and automation. It allows operators to dynamically allocate resources and quickly deploy new services across the 5G Core and RAN components while improving network agility and reducing capital and operational expenses. NFV is key in supporting the diverse requirements of 5G, from extreme mobile broadband to massive IoT connectivity and ultra-reliable low-latency communications, facilitating the evolution to more programmable, software-driven networks.


2.Cloud and Virtualization Tools ecosystem

What is Virtualization?

  • Virtualization is a technology that allows one physical machine to behave like many separate machines.

  • Virtualization lets you use hardware more efficiently and isolate workloads.

What is Cloud?

  • Cloud is an on-demand delivery of compute, storage, networking, and applications over the internet.

  • Cloud was built using virtualization, but it adds automation, elasticity, self-service, APIs, scaling, multi-tenancy and more.

Cloud = Virtualization + Automation + Self Service + Scalability


Cloud vs Virtualization

Virtualization

Cloud

Splits one physical machine into VMs

Provides infrastructure + services over internet

Focus = resource utilization

Focus = scalability, automation, self-service

Managed manually or via hypervisor

Managed via dashboards/APIs/orchestrators

Limited elasticity

Highly elastic – scale in/out instantly

You provision hardware yourself

Provider delivers resources on demand

Short answer:

Virtualization is a building block.

Cloud is the entire city built using these buildings.

Where Do Containers Fit?

Virtualization runs virtual machines. Cloud runs VMs + containers + serverless workloads.

Containers(Docker, containerd, CRI-O) = lightweight and faster alternative to VMsKubernetes orchestrates containers → forming modern cloud-native systems.

  • VM  =  Heavy, includes OS, slow startup

  • Container = Lightweight, shares OS kernel, fast startup

The Below image shows a Cloud & Virtualization Tools Ecosystem, organized into layers.


ree

Think of it like building a house:

Layer

What it provides

Without it…

Infrastructure

The foundation (bare-metal, VMs, cloud)

No compute/storage/networking to run workloads

Container Runtime

Packaging & running applications in containers

Apps would run directly on OS without isolation

Orchestration

Managing and scaling many containers across nodes

You would manually start/stop each container

Service Mesh & Management

Traffic control, packaging, security, monitoring

Large microservice systems would become unmanageable

1. Infrastructure Layer

Bare Metal → Virtual Machines → Cloud ProvidersThis is the base layer where compute, memory, and networking come from.

Tool

What it is

Why it exists / Use cases

OpenStack

Opensource private cloud IaaS platform

Create VMs, networks, storage within your data center

VMware

Hypervisor-based virtualization

Enterprise-grade VM management, vSphere is widely used in telco

KVM/Xen

Open-source hypervisors

Lightweight virtualization, basis for many telco clouds

AWS / Azure / GCP

Public cloud providers

On-demand compute, managed Kubernetes, global scale

Proxmox

Open-source virtualization platform

Small-scale labs, home labs, edge deployments

 

2. Container Runtime Layer

Where applications are packaged & executed as containers.

Tool

What it does

Key points

Docker

Most popular container runtime + image builder

Easy images, builds, registries — widely adopted

containerd

Lightweight runtime used under Docker

Kubernetes default runtime in many distros

CRI-O

Kubernetes-native runtime

Built specifically for Kubernetes workloads

Podman

Daemonless Docker-compatible runtime

Rootless, more secure on hosts

 

3. Orchestration Layer

Automates running hundreds or thousands of containers across nodes.

Tool

Description

Why it matters

Kubernetes (K8s)

Self-healing, autoscaling orchestration platform

Industry standard for CNFs (Cloud-native Network Functions)

Docker Swarm

Simple clustering for Docker containers

Easier but less powerful than Kubernetes

Nomad

Lightweight orchestrator by HashiCorp

Multi-runtime (VMs, containers, binaries) flexibility

Apache Mesos

Resource scheduler for large scale compute farms

Used before Kubernetes dominated

 

4. Service Mesh & Management Tools Layer

Enhances communication, lifecycle management, security between microservices.

Tool

Category

Purpose

Helm

Kubernetes Package Manager

Install/Upgrade CNFs like software packages

Istio

Service Mesh

Traffic control, security, metrics, canary rollout

Linkerd

Service Mesh (lightweight)

Simpler alternative to Istio

Consul

Service Discovery + KV Store

Dynamic service registration + config management

Envoy

Service Proxy

Forms data-plane of service meshes (sidecar proxy)

Kustomize

Config Management

Patch/overlay Kubernetes YAML without templates

 

Why Telcos Moved to Cloud + Virtualization

Benefit

Impact for Telco

Scalability

Scale UPF/AMF/SMF during busy hours

High Availability

Zero-downtime upgrade using Helm/K8s

Resource Optimization

Run multiple CNFs on same nodes

Faster Deployment

Deploy new core slice in minutes

Multi-Site Edge Cloud

Low latency RAN workloads near users

3.5G End-to-End Cloud & NFV Architecture

5G end-to-end deployments virtualize the core (5GC), RAN (CU/DU/RU), and RIC (Near-RT/Non-RT) on telco cloud stacks, disaggregating functions into CNFs/VNFs for scalability.


Core Components Explained

  1. 5G Core (5GC): Virtualized AMF/SMF/UPF microservices handle session management and user plane; deployed as CNFs on Kubernetes for auto-scaling during peaks.​

ree

 

  1. RAN (CU/DU/RU):

CU-CP: Control plane (RRC/PDCP) in central DCs.

CU-UP: User plane (SDAP, PDCP/RLC) near DUs.

DU: Real-time MAC/PHY on edge with DPDK.

RU: Hardware radio with eCPRI to DU.​


  1. RIC (RAN Intelligent Controller):

Non-RT RIC: rApps (>1s latency) for AI/ML Optimizations, Policy decisions (e.g., QoS) in central cloud via DMaaS (Data Management).

Near-RT RIC: xApps for beamforming/handover optimization (<10ms latency).​

ree

4.Description about Cloud/Virtualization Tools


Kubernetes: CU, DU, 5GC, Edge UPF

Kubernetes (K8s) orchestrates containerized network functions (CNFs) across clusters, automating deployment, scaling, and healing for 5G CU (control/user plane), DU (baseband), 5GC (AMF/SMF/UPF), and edge UPF (user plane function near users). It uses pods for microservices, node selectors for edge placement, and Horizontal Pod Autoscaler (HPA) for traffic spikes; essential for O-RAN multi-vendor interoperability.​

Docker/Container: All Containerized NFs

Docker packages NFs into portable OCI-compliant images (e.g., quay.io/5gc-amf:latest), enabling consistent deployment of all CNFs across dev/test/prod. Containers provide isolation, fast startup (<1s vs. VM minutes), and layering for updates; CRI-O/Podman run these images in K8s for 5G, caching locally to reduce pull times during DU restarts.​

OpenStack: 5GC (VM-based), CU (Legacy)

OpenStack manages infrastructure-as-a-service (IaaS) for VM-based VNFs, provisioning compute/storage/networks for legacy 5GC (e.g., full AMF VM) and CU in brownfield networks. It creates flavors (e.g., highcpu-du), Neutron networks for Xn/F1, and Heat templates for orchestration; bridges VMs to CNF transition in hybrid telco clouds.​

Helm: Package Management for All K8s Workloads

Helm acts as a package manager for Kubernetes, bundling charts (YAML templates + values) for repeatable 5G deployments like helm install open5gs/5gc --set replicas=3. It simplifies versioning, dependencies (e.g., Istio), and upgrades (helm upgrade), reducing manual kubectl apply errors for complex CU/DU/RIC stacks.​

Istio/Linkerd: Service Mesh for 5GC, CU Inter-NF

Istio and Linkerd provide service mesh overlays with sidecar proxies (Envoy) for secure communication between 5GC microservices (N4 interface) and CU components (E1/F1). Features include mTLS encryption, traffic splitting (blue-green deploys), observability (Prometheus metrics), and circuit breaking; Linkerd is lighter for edge while Istio offers advanced policy.​

KVM/Bare Metal: DU (Performance-Critical)

KVM (Kernel-based Virtual Machine) hypervisor virtualizes on bare metal x86 servers for latency-sensitive DU (MAC/PHY processing), using SR-IOV for direct NIC access and NUMA pinning (numactl) for <100μs jitter. Bare metal bypasses hypervisor overhead entirely for ultimate DU performance in 25Gbps eCPRI fronthaul; DPDK accelerates packets.​

K3s/MicroK8s: Lightweight K8s for Edge/DU

K3s (lightweight Kubernetes by Rancher) and MicroK8s (Canonical's single-node K8s) run on resource-constrained edge servers for DU/UPF, embedding etcd and CRI-O with <500MB footprint. Ideal for far-edge cell sites; k3s install deploys in seconds, supporting air-gapped ops with embedded containerd—no full K8s master complexity.


Summary of Tools

ree

Tool

Primary 5G Use

Key Benefit

Kubernetes

Orchestration

Auto-scaling ​

Docker

Packaging

Portability ​

OpenStack

IaaS VMs

Multi-tenancy ​

Helm

K8s Packages

Templating ​

Istio/Linkerd

Traffic Mgmt

mTLS/Resilience ​

KVM/Bare Metal

DU Compute

Low Latency ​

K3s/MicroK8s

Edge K8s

Lightweight ​

5.Telco Cloud Application deployment Workflow

7-Step Deployment Pipeline:

  1. Development Phase - Developers write code for 5G network functions, create Dockerfiles for containerization, and define Helm charts for deployment configuration

  2. Source Control - Code is pushed to Git repositories (GitHub/GitLab) with version tagging and code review processes

  3. CI Pipeline (Build) - Automated testing, security scanning (SAST/SCA), Docker image building, and vulnerability scanning happen automatically when code is committed

  4. Container Registry - Built images are stored in registries (Harbor, Quay, ECR) with proper versioning, signing, and access controls

  5. CD Pipeline (Orchestration) - ArgoCD or Flux automatically deploys applications using Helm charts to Kubernetes clusters with rolling updates and health checks

  6. Infrastructure - Applications run on Kubernetes clusters across different locations (central cloud, edge, RAN sites) on OpenStack VMs or public clouds

  7. Monitoring & Logging - Continuous monitoring with Prometheus/Grafana, log aggregation with ELK/Loki, and alerting systems provide observability

Feedback Loop - Monitoring data feeds back to development for continuous improvement.


ree

6.Cost Savings for Telco Providers:

1. Infrastructure Cost Reduction (30-50%)

  • Hardware Consolidation: Run multiple VNFs on shared infrastructure instead of dedicated hardware per function

  • Cloud Economics: Use COTS (Commercial Off-The-Shelf) servers instead of expensive proprietary hardware

  • Dynamic Resource Allocation: Scale resources up/down based on demand

2. Operational Efficiency (40-60% OpEx reduction)

  • Automation: CI/CD eliminates manual deployment (reduces human errors and time)

  • Self-Service: Developers can deploy without waiting for operations teams

  • Faster Time-to-Market: Deploy new features in hours instead of months

3. Multi-Vendor Flexibility

  • No Vendor Lock-in: Use best-of-breed components from different vendors

  • Negotiation Power: Switch vendors more easily, reducing licensing costs

  • Open Source: Leverage free open-source tools (Kubernetes, Prometheus, etc.)

4. Resource Optimization

  • Better Utilization: Increase server utilization from 20-30% to 60-80%

  • Energy Savings: Fewer physical servers = lower power and cooling costs

  • Space Savings: Reduced data center footprint

5. Faster Innovation

  • Rapid Testing: Test new features in isolated environments

  • A/B Testing: Run multiple versions simultaneously

  • Rollback Capability: Quick recovery from failures

6. Centralized Management

  • Single Control Plane: Manage all sites from one location

  • Consistent Deployments: Same process across all locations

  • Reduced Site Visits: Remote management reduces truck rolls


7.Real-World Case Study:


1) Verizon/VMware Telco Cloud

Verizon deployed vRAN on VMware Telco Cloud (KVM/OpenStack base) with Kubernetes/CRI-O for CU/DU across 20k+ sites: CU-CP on central K8s, DU CNFs on edge with Istio for F1 resilience, achieving 40% cost reduction vs. proprietary RAN. RUs used Nokia hardware with eCPRI to vDU; automated scaling handled 5G traffic spikes via Prometheus-tuned HPA.​

Method of Procedure (MOP) for Design & Deployment

  1. Capacity Planning: Model traffic (e.g., 1Gbps/DU); allocate flavors (32C/128G for DU) via OpenStack.​

  2. Cluster Provisioning: 

    • kubeadm init --pod-network-cidr=x.x.x.x/16;

    • install CRI-O (crio install)

    • label nodes (kubectl label node edge telco-du=true).​

  3. CNF Deployment: 

    • Helm install CU/DU charts (helm install cu-cp o-ran-chart);

    • configure SR-IOV (kubectl apply sriovdp.yaml).​

  4. Fronthaul Setup: 

    • Bind RUs to DU via eCPRI sync (PTP/IEEE 1588);

    • apply Istio VirtualService for F1/E1.​

  5. Validation: 

    • RIC testing (xApp deploy for O-RAN);

    • perf test (tmux iperf3 -c ru-sim);

    • HA failover (kubectl drain).

    • Ensures 99.999% uptime.


2) Rakuten Mobile Japan

Rakuten deployed Open RAN on OpenStack/Kubernetes (CRI-O) for 5GC (free5GC CNFs), vRAN (CU/DU on edge K8s), and Near-RT RIC (xApps for mobility). Covering 96% Japan by 2024, it achieved $1B savings using white-box hardware; Istio secured F1 interfaces, scaling to 10M subs with <5ms E2E latency. Non-RT RIC policies via rApps boosted throughput 25%.​

Method of Procedure (MOP) for E2E Deployment

  • Infrastructure Provisioning: 

    • Install KVM (modprobe kvm);

    • deploy OpenStack (devstack all-in-one for lab, Kolla-Ansible for prod);

    • create flavors (openstack flavor create --ram 128 du-edge).​

  • Kubernetes Cluster Bootstrap: 

    • kubeadm init --pod-network-cidr=x.x.x.x/16 --cri-socket unix:///var/run/crio/crio.sock (CIDR=Flannel pods, CRI=CRI-O);

    • taint edge nodes (kubectl taint nodes edge sched=du).

    • Join workers (kubeadm join).​

  • 5GC CNF Deployment: 

    • helm repo add open5gs;

    • helm install 5gc open5gs/5gc --set upf.image.repository=quay.io/upf:

    • latest (pulls Docker images).​

  • RAN (CU/DU/RU) Rollout:

    • SR-IOV: kubectl apply -f sriov-network-device-plugin.yaml.

    • CU/DU: kubectl apply -f cu-cp.yaml -n ran-ns (--namespace isolates); kubectl apply -f du-cnf.yaml --dry-run=client (validates first).​

    • RU Connect: Configure eCPRI PTP (chronyd.conf); Istio istioctl install --set profile=default -y.​

  • RIC Integration: 

    • Non-RT: OpenStack VM for rApps.

    • Near-RT: kubectl apply -f near-ric.yaml; E2 SM for xApps.​

  • Validation: 

    • kubectl get pods -n ran; traffic test (iperf3 -c upf-svc);

    • scale (kubectl autoscale du --min=3 --max=10 --cpu-percent=70).

    • Monitor (istioctl dashboard proxy-istio).

    • Ensures greenfield 5G launch in 3-6 months


8.Common telco ops tasks + commands

  • Check pod/containers status

    • kubectl get pods -n telco-namespace

    • kubectl describe pod <pod> -n telco-namespace

    • kubectl logs <pod> -c containername -n telco-namespace

  • Scale a component (e.g., UPF dataplane)

    • kubectl scale deployment upf-deployment --replicas=5 -n telco

  • Rolling update via Helm

    • # change image tag in values.yaml then

    • helm upgrade core-network ./core-chart -n telco --reuse-values

  • Istio traffic shaping

    • Canary: modify VirtualService weights to shift traffic.

    • Fault injection: test resilience by injecting HTTP 500 for a percentage (for testing only).

  • Debugging network issues

    • kubectl exec -it <pod> -- /bin/bash

    • # inside: curl other services, netstat, ip route

  • Persistent storage (e.g., for HSS/DB)

    • Use PersistentVolume and PersistentVolumeClaim in k8s, or OpenStack Cinder volumes for VMs.

 

9.Summary

1.      Cloud and virtualization tools let operators deploy networks rapidly without waiting for physical hardware.

2.      They enable on-demand provisioning of compute, storage, and network functions as software rather than appliances.

3.      New 4G/5G sites or Core instances can be spun up in minutes using automated orchestration instead of days of manual setup.

4.      Containerization and Kubernetes make network functions scalable, self-healing, and easy to upgrade with zero downtime. Telcos can expand capacity instantly by adding more VNFs/CNFs rather than installing new boxes.

5.      Centralized automation reduces rollout cost, accelerates time-to-market, and supports coverage expansion at scale.

Overall, these tools transform network buildout from hardware-driven to software-driven — faster, flexible, and future-ready.

 

10.References

1.       3GPP TS 23.501 System Architecture for the 5G System (5GS)

2.       3GPP TS 28.500 — Management concept, architecture and requirements for mobile networks that include virtualized network functions (VNFs).

3.       O‑RAN Architecture Description (version v14.0)

Comments


 

bottom of page