Arch End To End Low Latency Wireless Networks
Arch End To End Low Latency Wireless Networks
Cisco public
In Voice over IP (VoIP), 150 milliseconds (ms) of latency in one direction is not noticeable by users, and
therefore is perfectly acceptable. With applications like collaboration (Webex, Microsoft Teams), Office 365,
messaging, and 3D collaboration with augmented and Virtual Reality (AR/VR), sub-50-ms bidirectional
response times are desired. Similarly, if you’re using wireless connections to run an Autonomous Mobile Robot
(AMR) or an Automated Guided Vehicle (AGV) in a factory, sub-20-ms response times in a high-throughput
network are required, while some process control loop traffic requires 10 ms or less.
In network design, achieving end-to-end IP latency to meet the needs of specific applications requires an
understanding of the data flow, technologies, and topology of the overall network architecture. However, quite
often specific Layer 1 and Layer 2 technology developers advertise only their own latency values while
neglecting to factor in end-to-end IP latency, which is the sum of all Layer 1 and Layer 2 infrastructures that IP
flows traverse.
End-to-end IP latency is usually calculated in one direction: from the wireless device to the wireless network, IP
transport network, and application server (Figure 1). RTT is the calculation of bidirectional latency (e.g., the time
required for a network ping). In addition, the processing time required before a response is sent back must also
be considered in calculating overall latency.
Figure 1.
End-to-end IP latency
3GPP efforts related to 5G latency have focused on reducing and controlling the time spent by traffic over the
radio network. While latency target numbers were defined dependent on the wireless service type (such as
eMBB ~4 ms or URLLC ~1 ms) for a specific data flow throughput, delays and throughput loss from various
network functions must be factored in when designing networks, as actual latency will be higher in typical
operations.
Like 3GPP, the IEEE 802.11 working group looks at improving the deterministic aspects of the Wi-Fi protocol. A
recent study of deterministic Wi-Fi evaluated the latency of existing Wi-Fi versions and how enhancements
could be made available on the ongoing IEEE 802.11be or Wi-Fi 7 specifications (Figure 2). This evolution of
enhancements, which also demonstrates the need for collaboration between vendors, is designed to meet new
requirements by industrial applications (for example, for AMR and AR/VR applications) with the correct network
design of the Wi-Fi segment.
Each wireless technology standard tries to improve the radio characteristics of the solution, but each represents
only one of many enhancements to be considered in a network design.
Figure 2.
Evolution of Wi-Fi latency
Achieving lower RTT latency is simplified with the closer proximity of applications hosting the wireless devices.
In the calculator in Figure 3, RTT through fiber links is estimated, which helps us understand the additional delay
added through typical fiber optic paths when communicating with cloud or data center applications. If a network
design targets less than 10 ms of RTT latency through the IP transport infrastructure, the selection of a
managed cloud provider is required, one that offers data center services located within 200 kilometers of the
device’s location. The example in Figure 3 is based on France, which mandates at least 10 regional data
centers for such coverage, without considering data and application movement around the cloud.
This example perfectly illustrates why achieving lower end-to-end IP latency (such as 5 ms) for edge
computing is a mandatory component of application hosting that must be collocated within campuses or
regional data centers.
Figure 3.
5G Cloud topology with a sub-10-ms RTT (Source: OHVCloud)
Here is what low-latency architectures look like in private 5G, Wi-Fi 6/6E, and Cisco URWB deployments.
These advanced 5G solutions include optimizations at every step of the radio hardware and uplink and downlink
transmission processes. New radio features address low-latency communications, allowing for a variable
Transmission Time Interval (TTI) that can scale from 1 ms down to ~140 microseconds, depending on whether
spectral efficiency in eMBB or low latency in URLLC is the main goal.
In a private 5G topology within a factory, as shown in Figure 4, an industrial robot connected to the Radio
Access Network (RAN) is shown at lower left. Once IP packets exit the robot, they are encapsulated inside a
tunnel and the intermediate devices don’t see IP packets but instead recognize the MAC layer encapsulation.
Figure 4.
5G data flow with an open RAN architecture
The IP packets exit the tunnel at the User Plane Function (UPF). The UPF is the interconnect point between the
mobile infrastructure and the data network (that is, encapsulation and decapsulation of General Packet Radio
Service [GPRS] Tunneling Protocol for the user plane). The packets are then routed to the application, either
locally or to a data center or an internet cloud service. One benefit is that the UPF can be distributed, allowing a
central UPF to serve as anchor for the overall network, while the local UPF may be deployed on a site to reduce
the path to local applications.
Again, the location of cloud providers or company data centers has a major impact on latency. The closer they
are located to the end devices, the lower the latency. Latency results from a full path through all these devices.
Figure 5.
Wi-Fi low-latency architecture
Wi-Fi and 5G use different types of encapsulations, but IP packets in a W-Fi network similarly move from
access points to the wireless radio network and through tunnels to a Wireless LAN Controller (WLC). In a
network design, the end-to-end IP path to applications in a cloud or data center from the WLC is identical to
that in the 5G use case, since it is independent from the selected wireless technology. Similarly, for private 5G
deployments, if low latency is required for an application, a Wi-Fi WLC for the application server path should be
designed to yield a path that is as short as possible.
One main benefit of Wi-Fi architectures is the ability to do local switching without reaching a WLC (Figure 5),
another factor that may contribute to lower latency in a design.
Cisco URWB technology provides low-latency, highly reliable, long-range, high-bandwidth connections that
can handle endpoints moving at high speeds with zero-delay handoffs. Operating in unlicensed frequencies, the
Cisco URWB segment requires an appropriate design to control latency and fast handover in less than 5 ms,
while the end-to-end IP infrastructure beginning at the Cisco URWB gateway is like Wi-Fi and 5G topologies.
Recent enhancements were made to Cisco URWB to deliver uninterrupted connectivity to fast-moving devices
by sending high-priority packets via redundant paths. Cisco patented Multipath Operations (MPO) technology
can duplicate protected traffic up to 8x and avoid common paths, and it works alongside hardware availability
for lower latency and higher availability, limiting interference and hardware failures.
A Cisco URWB topology is shown in Figure 6. The onboard layer at the bottom represents a vehicle that
contains a radio connected to a base station. Data from the vehicle is encapsulated on a centralized controller
and then decapsulated as IP packets on the IP network. Again, lower levels of latency can be realized per
application by moving the data center or cloud provider closer to the end device.
Figure 6.
Cisco Ultra Reliable Wireless Backhaul topology
Whether on a factory floor, in a university research center, on a high-speed train, or in a branch office, the
better the network design, the lower the probable service latency.