How to achieve low latency audio/video streaming over IP network?

Similar documents
Assuring Media Quality in IP Video Networks. Jim Welch IneoQuest Technologies

IP Video Network Gateway Solutions

Important Encoder Settings for Your Live Stream

CS 218 F Nov 3 lecture: Streaming video/audio Adaptive encoding (eg, layered encoding) TCP friendliness. References:

Chapter 9. Multimedia Networking. Computer Networking: A Top Down Approach

Megapixel Networking 101. Why Megapixel?

WHITE PAPER. Atlona OmniStream: Truly Converged, Networked AV. US International

Scalable Video Coding

SMPTE ST In Real World Applications. Paul Macklin (Vimond) and Alexander Sandstrom (Net Insight)

INTRODUCTORY Q&A AMX SVSI NETWORKED AV

Chapter 28. Multimedia

Page 1. Outline / Computer Networking : 1 st Generation Commercial PC/Packet Video Technologies

AV OVER IP DEMYSTIFIED

Digital Asset Management 5. Streaming multimedia

WHITE PAPER. Atlona OmniStream: Truly Converged, Networked AV

Recommended Readings

Contents. Introduction. Diagram. Software. Key Features. Functionalities. Specifi cations. Typical Applications

Perfect Video over Any Network. State-of-the-art Technology for Live Video Comunications

Internet Video Delivery. Professor Hui Zhang

Multimedia Networking

Technology Overview. Gallery SIENNA London, England T

15: OS Scheduling and Buffering

Multimedia Networking

Image and Video Coding I: Fundamentals

Why Shaping Traffic at the Sender is Important. By Chuck Meyer, CTO, Production December 2017

JPEG K UHD Video over IP Decoder with KVM

Datasheet EdgeVision Multichannel Quality of Experience Monitoring

Multimedia: video ... frame i+1

CODING METHOD FOR EMBEDDING AUDIO IN VIDEO STREAM. Harri Sorokin, Jari Koivusaari, Moncef Gabbouj, and Jarmo Takala

H.264 AVC 4k Decoder V.1.0, 2014

CN1047 INTRODUCTION TO COMPUTER NETWORKING CHAPTER 6 OSI MODEL TRANSPORT LAYER

RECOMMENDATION ITU-R BT.1720 *

INTRODUCTION TO JPEG XS THE NEW LOW COMPLEXITY CODEC STANDARD FOR PROFESSIONAL VIDEO PRODUCTION

Model: LT-122-PCIE For PCI Express

irtc: Live Broadcasting

Introduction to LAN/WAN. Application Layer 4

HDMI/HD-SDI HEVC/H.264 IPTV

Achieving Low-Latency Streaming At Scale

CS 5520/ECE 5590NA: Network Architecture I Spring Lecture 13: UDP and TCP

JPEG K UHD Video over IP Decoder with KVM

ELEC 691X/498X Broadcast Signal Transmission Winter 2018

Outline. QoS routing in ad-hoc networks. Real-time traffic support. Classification of QoS approaches. QoS design choices

Sonifex Pro Audio Streamers Frequently Asked Questions (FAQ)

JPEG K UHD Video over IP Encoder with KVM

Testing Video over IP Product and Services

An Efficient Verification Framework for Audio/Video Interface Protocols

MA5400 IP Video Gateway. Introduction. Summary of Features

MITIGATING THE EFFECT OF PACKET LOSSES ON REAL-TIME VIDEO STREAMING USING PSNR AS VIDEO QUALITY ASSESSMENT METRIC ABSTRACT

OSI Layer OSI Name Units Implementation Description 7 Application Data PCs Network services such as file, print,

Networking Applications

IO [io] MAYAH. IO [io] Audio Video Codec Systems

Application and Desktop Sharing. Omer Boyaci November 1, 2007

JPEG K UHD Video over IP Encoder with KVM NMX-ENC-N2151 (FGN2151-SA), Stand Alone NMX-ENC-N2151-C (FGN2151-CD), Card

Image and video processing

Streaming (Multi)media

DM-TXRX-100-STR HD Streaming Transmitter/Receiver. Supplemental Guide Crestron Electronics, Inc.

Request for Comments: 4425 Category: Standards Track February 2006

ENVISION INFINITE CONNECTIVITY.

MISB EG Motion Imagery Standards Board Engineering Guideline. 24 April Delivery of Low Bandwidth Motion Imagery. 1 Scope.

Digital Media Capabilities of the Modero X Series Touch Panels

CS640: Introduction to Computer Networks. Application Classes. Application Classes (more) 11/20/2007

ECE 650 Systems Programming & Engineering. Spring 2018

CS 528 Mobile and Ubiquitous Computing Lecture 4a: Playing Sound and Video Emmanuel Agu

IOCAST video transmission solutions

Chapter 7 Multimedia Networking

Asynchronous Transfer Mode (ATM) ATM concepts

H.264 ENCODER AND DECODER

Fundamentals of Video Compression. Video Compression

A Proposed Time-Stamped Delay Factor (TS-DF) algorithm for measuring Network Jitter on RTP Streams

The Frozen Mountain irtc White Paper Series

Optimizing IP Networks for Acquisition. Presented by Henry Quintana, Director of Solutions TVU Networks

CSC 4900 Computer Networks: Multimedia Applications

ECS 152A Computer Networks Instructor: Liu. Name: Student ID #: Final Exam: March 17, 2005

Cobalt Digital Inc Galen Drive Champaign, IL USA

VoIP. ALLPPT.com _ Free PowerPoint Templates, Diagrams and Charts

Real-Time Protocol (RTP)

Full HD HEVC(H.265)/H.264 Hardware IPTV Encoder Model: MagicBox HD4 series MagicBox HD401: Single channel HDMI/AV, HDMI/VGA/YPbPr/AV, HDSDI input

Kommunikationssysteme [KS]

Multimedia Storage Servers

Multimedia in the Internet

Transport protocols Introduction

Designing and Provisioning IP Contribution Networks

Ideal Systems integrate Bluefish444 supported workflows

MODERN /CONTROLROOMS THE FINDING A SOLUTION THAT WORKS FOR THE WAY YOU WORK

CSCD 433/533 Advanced Networks Fall Lecture 14 RTSP and Transport Protocols/ RTP

ACCESS 4.0 Addendum. Product Manual

Data Representation and Networking

HDMI/HD-SDI/VGA H.264/H.256 HEVC

ST2110 and High Bitrate Media Transport over IP Networks

Request for Comments: 5109 December 2007 Obsoletes: 2733, 3009 Category: Standards Track. RTP Payload Format for Generic Forward Error Correction

CSC 401 Data and Computer Communications Networks

Multimedia Protocols. Foreleser: Carsten Griwodz Mai INF-3190: Multimedia Protocols

Q-SYS NV-32-H Preliminary Specifications Q-SYS NV-32-H. Features. Native integration and control for the Q-SYS Ecosystem

Chapter 7. The Transport Layer

Transporting audio-video. over the Internet

MODERN /CONTROLROOMS THE FIND A SOLUTION THAT WORKS FOR THE WAY YOU WORK

The Pros and Cons of JPEG 2000 for Video Archiving. Katty Van Mele November, 2010

Module 7 VIDEO CODING AND MOTION ESTIMATION

Model: LT-101-PCI-VC1

Multimedia networked applications: standards, protocols and research trends

Transcription:

February 2018 How to achieve low latency audio/video streaming over IP network? Jean-Marie Cloquet, Video Division Director, Silex Inside Gregory Baudet, Marketing Manager, Silex Inside Standard audio video interfaces such as HDMI and Display Port are well suited for short range connectivity of multimedia equipment. However, larger AV (Audio/Video) installations with multiple sources and displays, or installations spreading over several physical locations need more evolved connectivity. For those AV installations, it has become obvious that IP networks is the most standard and future proof way of transporting the signals. Although IP-based networks enable the flexibility and scalability required by many applications, special care should be taken to keep the latency of the system sufficiently low for real-time, live use cases. This whitepaper will first define the latency for audio/video transport. It will highlight the usual architecture challenges of an AV over IP transmitter/receiver. A deeper analysis is provided regarding the video compression which is often mistakenly considered as adding too much latency. Eventually, actual latency measurements of the Viper 4K HDMI to IP transmitter/receiver will be presented. How is the latency defined for an audio/video transmission? The latency of a system is the delay between the instant a sample enters the system and the instant it leaves the system. In an AV over IP system, it translates into the delay between the first pixel of a video frame entering the transmitter through the video input and the first pixel of the same video frame going out of the receiver on the video output. The latency is naturally defined in seconds usually in the range of milliseconds for a real-time audio-video system. Video experts also define the latency as the corresponding part of the video stream during that time. The latency is then described as a number of frames or lines of a video stream. In this case, the actual time varies depending on the frame rate of the video as shown in the table below. Latency of 1 frame (ms) Latency of 1 line (ms) 720p 50fps 20 0.0278 1080p 30fps 33.3 0.0309 UHD 60fps 16.7 0.0078 Table 1 Latency equivalence of one frame and one line of video signal for different video formats. www.silexinside.com 1

This definition is very convenient for some image processing algorithm where the added latency is, for example one frame, whatever is the frame rate. There isn t a unique definition of what should be the latency of an AV over IP system. Low Latency, Ultra- Low Latency, or even Zero Latency are commonly used terms to indicate that the latency is good enough for the intended application and end user expectation. Applications that involve the interaction of the user (like Meeting Presentation, KVM or live events) are usually the most critical in terms of latency. Some users will be more sensitive to the latency than others, but keeping the latency below 30 ms is usually accepted. Some applications benefit from even lower latency for a seamless user experience. Architecture challenges of a low latency AV over IP system Special care should be taken to the architecture of the transmitter and receiver in order to achieve low latency AV over IP. The latency of the system directly comes from the buffering of the video/audio at the different processing stages. This buffering is necessary to enable some features, but should be kept to a minimum as described later. Due to the high bandwidth of the video signal and the latency constraint, it is essential to use dedicated hardware processing from the video input to the IP network. Purely software-based solutions will inevitably increase the latency because of the memory transfers and the CPU load. Although having dedicated hardware support for the video stream is mandatory, it needs to remain configurable and flexible. For this reason, the software running on a CPU takes care of all the non-real-time tasks. The following diagram shows the basic processing stages of an AV over IP platform. Figure 1 Block diagram of the architecture of a typical AV over IP transmitter or receiver A typical transmitter takes the video from its input, sends it through some video processing, video encoding and network processing before it outputs the stream on the IP network. A receiver does similar operations in reverse order. Each processing step can potentially add latency to the complete system and deserves a deeper analysis. www.silexinside.com 2

Video input and output Receiving and transmitting on the video interfaces, such as HDMI, Display Port, SDI does not add latency to the system. A few frames may be discarded at startup during the initialization process of the input and output stages, but this does not add latency. When the video content is protected with HDCP, there is an additional authentication phase that takes place when the cable is plugged in. After this authentication phase, the video can be encrypted/decrypted on-the-fly without any additional buffering, therefore without adding latency. Another important aspect is that the video input and output of the AV over IP systems are located on two different devices connected together via the IP network. One of the challenges is that the video input of the transmitter board needs to run at the exact same frequency as the video output of the receiver board. If it wouldn t be the case, the receiver would have too many or not enough data to output on the video link making it quickly unstable. This issue is sometimes solved with a frame buffer at the receiver that can drop or repeat a frame when necessary, but this adds a frame of latency. The best approach is to implement a clock recovery mechanism over the network that will replicate the video clock of the transmitter at the receiver, guaranteeing synchronized operation. Video processing Transmitters and receivers often include video processing functionalities. It may include among others scaling, chroma up/down sampling, color conversion, frame rate conversion and image filtering. Most of the video processing functionalities are described by a filter function. Each filter requires a certain amount of data to be buffered during the calculations, adding up to the total latency. If a filter uses pixels of a single line, the latency is negligible. The impact is more important if a filter uses a large part of the frame or even pixels from previous frame(s). Video compression Video compression is used to reduce the bitrate of the video. In the case of video transport over IP, reducing the bitrate has a direct positive effect on the network infrastructure costs. It also enables more video streams to be transported on a specific network installation without congestion. The following table gives an overview of the bandwidth of the raw video (uncompressed), together with the minimum compression ratio required to fit in 1G or 10G Ethernet. Video source 3G-SDI 1920x1080, 60fps, 4:2:2, 10-bit Blu-Ray UHD 3840x2160, 60fps, 4:2:0, 10-bit HDMI 2.0 4096x2160, 60fps, 4:4:4, 8-bit Raw Bandwidth Compression Ratio 1Gb Eth Compression Ratio 10Gb Eth 2.5 Gb/s 3:1 NA 7.5 Gb/s 9:1 NA 12.7 Gb/s 15:1 1.5:1 Table 2 Example of compression ratio for several video formats. The raw bandwidth does not take into account the vertical blanking of the video signal. It is often said that compression adds a huge latency to a system. This is simply not true if the right codec is www.silexinside.com 3

selected. When choosing a compression algorithm, there are a lot of aspects to take into account like the compression ratio to achieve, the quality expectation, the complexity in hardware or software, the interoperability with other equipment and of course the latency. Each application has a different set of requirements that will lead to one or multiple possible codecs. The video compression topic is further explained in the next chapter. Network transport Before the audio and video data can be sent over the IP network, it needs to be encapsulated in several protocols. The audio and video is usually transported in RTP packets that are themselves encapsulated in UDP/IP packets. UDP protocol is used for the real-time transport as it allows broadcast/multicast. Moreover a connection-oriented protocol with packet acknowledgement and retransmission such as TCP would not work for a real-time and low latency transport. Each IP packet is encapsulated in an Ethernet frame. The payload of the Ethernet frame is limited to maximum 1500 bytes on general purpose networks. For this reason, each video frame is divided into many small packets for the transmission. The receiver reconstructs the video frames by concatenating the data of all the packets. It is essential to process the packets in real-time to maintain the low latency, and not accumulate them in a buffer, for example, until a complete video frame would be ready. Using a hardware in-line packet engine, it is very easy to execute these tasks with negligible latency. The transported audio uses much less bandwidth than the video and special care should be taken for the encapsulation in network packets. Indeed, audio samples are grouped in small amount of samples to avoid adding delay at the encapsulation. It is also necessary to keep a relatively small packet time for the clock recovery mechanism to be reliable. The network infrastructure itself adds its own latency, but this is usually very limited on a local area network (less than a millisecond). Larger networks can also introduce some jitter at the packet level which needs to be properly handled at the receiver side. In practice, the receiver has a small network packet buffer to compensate for the jitter and the granularity at which the video decoding can be done. This buffer should be configured to the minimum that guarantees the reliable operation of an installation. How to achieve sub-frame latency with video compression? There are many different video codecs for different purposes. Selecting the right video codec is always a compromise between the latency, compression ratio and quality. It is not possible to score well in all criteria at the same time. As an example, a codec like h264/h265 used for the video distribution over Internet is optimized to achieve the best compression ratio and image quality at the expense of high latency. Neighboring pixels need to be involved in order to increase the efficiency of the compression. The pixels can be spatial neighbors (from the same frame), or temporal neighbors (from other previous or next frames). Most advanced video codecs in terms of compression ratio (like h.264/265) are called inter-frame codecs. They take advantage of this principle by using several frames before and after the current frame to encode it. This of course induces several frames of latency. In general, the codec latency is caused by the fact that future pixels are involved in the encoding of the current pixels. Some codecs also require several www.silexinside.com 4

passes with complex calculations that can also increase the latency depending on the hardware/software implementation used. Another aspect that affects the latency of the codec is the rate allocation mechanism that is used. The rate control of the encoder regulates the amount of compressed data to achieve the requested target bit rate on average. For a codec to be low latency, it is important that the bitrate is constant (CBR) on a small time window. The time window used for averaging the bitrate is important. For example, a codec could produce a stream at a bitrate that would be constant when averaged over 5 frames, but not necessarily constant when averaged over 1 frame. As the compressed stream is transported over a channel with limited bandwidth capacity, it is then required to use additional buffering and latency to smooth the transmission. For this reason, a very low latency codec will generate a constant bitrate output when averaged over a few video lines. Latency measurements of Viper 4K AV over IP This section shows a practical measurement of a sub-frame latency solution for AV over IP. The equipment used for the measurement is the Viper OEM board from Silex Inside running the VC-2 HQ codec. Figure 2 Viper hardware board with HDMI, analog audio and Ethernet interfaces. Viper low latency architecture The architecture of the Viper transmitter and receiver has been designed taking into account the principles that were described in the previous sections. Neither the transmitter nor the receiver stores any significant amount of data (such as full video frame) during the conversion from HDMI to IP and vice versa. The HDMI output clock of the receiver is synchronized to the HDMI input clock of the transmitter over the network in order to avoid any overrun or underrun of the receiver. On the video encoding side, the VC-2 HQ algorithm is used. VC-2 HQ is a SMPTE standard (SMPTE 2042) www.silexinside.com 5

ideally suited for low compression ratios (up to 10:1. VC-2 HQ has low complexity, and its line-based wavelet allows a latency of a few video lines only. The ultra-low latency of VC-2 is below the millisecond just like another simple video processing functionality. Another codec supported on the Viper boards is JPEG 2000. JPEG 2000 is a well-known JPEG standard that can achieve best quality for compression ratio up to 20:1. Full frame encoding/decoding has a latency of a few frames but it is permitted to encode/decode stripes (division of the frame in X lines) to bring the latency down to a few milliseconds. Both the VC-2 HQ and JPEG 2000 encoder produce a constant bitrate stream when averaged over a few video lines making it ideal for the transport over a fixed bandwidth network. This guarantees that no extra buffering is required during the transport and decoding of the stream. Latency measurement The latency of an AV over IP system is typically measured from the HDMI input of the transmitter to the HDMI output of the receiver. The measurement could be done with dedicated test equipment. However, for this whitepaper, we decided to use the embedded features of Viper to make the setup simpler. The transmitter and receiver are connected to each other via the IP network made up of a 1G switch. The subtlety of this test setup is that it uses the receiver to generate the audio and video HDMI stream that feeds the transmitter input. The latency measurement is done within the receiver device. The device generates a specific audio/video pattern that it is able to detect when it comes back after the transport over IP. The receiver calculates then the delay between the instant when the pattern is generated on the HDMI output, and the instant it is received in return over the Ethernet interface including the decoding and image processing in the receiver. Figure 3 Test setup used for the latency measurements. The test setup is used to measure the latency of the audio-video stream in different conditions. The video format can be freely modified, including resolution, chroma subsampling and bit-depth. The target bit-rate of the compressed video can also be changed in order to analyze the impact on the latency. The latency measurements are summarized in the following table: Test Video format Audio format IP Bandwidth End-to-end latency 1 1920x1080 60fps 4:2:2 10-bit 2ch PCM 16-bit 48kHz 900 Mbps 3.1 ms 2 1920x1080 60fps 4:2:2 10-bit 2ch PCM 16-bit 48kHz 300 Mbps 5.8 ms 3 3840x2160 30fps 4:4:4 8-bit 2ch PCM 24-bit 96kHz 900 Mbps 3.4 ms 4 3840x2160 60fps 4:2:0 8-bit 2ch PCM 24-bit 96kHz 900 Mbps 3.2 ms Table 3 Measurements of the total latency of Viper system with VC-2 HQ compression including receiver www.silexinside.com 6

and transmitter. The measurements clearly show consistent very low latency around 5 ms or less in all tested use cases. The latency slightly increases with a lower bitrate. This is due to the network stream buffering method implemented at the decoder that guarantees smooth streaming. The buffer size represents a larger video stream duration at lower bitrate. It can also be noticed that the frame rate has a very small impact on the latency, for example, comparing 30 fps vs 60 fps. This is possible because all the video processing and encoding data path only uses very few lines of the video. Conclusion This whitepaper has given an overview of the latency challenges when implementing an AV over IP solution. The latency comes from the many processing stages of the system. It is important to take the latency aspect into account from the beginning of the product design and architecture. The video codec, when properly selected, has a very low impact on the latency. When the system is well design, as shown in the Viper case, the latency can be as low as 5 ms and the 4K HDMI 2.0 video fits within a 1G Ethernet cable. This demonstrates the possibility to transport UHD AV content over 1G Ethernet with a seamless user experience. About Silex Inside Silex Inside is the leading provider of high-end image video and security technology. Silex Inside develops OEM solutions for audio/video compression and transport applications. Silex Inside also licenses a wide range of video compression technology, in the form of IP blocks for ASIC or FPGA. This technology has been adopted by leading companies in the broadcast, professional AV and defense market. Silex Inside also specializes in embedded security. Combined with video transport, this ensures secure video transmission and content protection solutions. For more information, visit us on www.silexinside.com, follow us on Twitter (@SilexInside), LinkedIn (SilexInside), YouTube (Silex Inside). www.silexinside.com 7