End-to-End Delay Analysis of Videoconferencing over Packet-Switched Networks

Size: px
Start display at page:

Download "End-to-End Delay Analysis of Videoconferencing over Packet-Switched Networks"

Transcription

1 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 8, NO. 4, AUGUST End-to-End Delay Analysis of Videoconferencing over Packet-Switched Networks Mario Baldi, Member, IEEE, and Yoram Ofek, Member, IEEE Abstract Videoconferencing is an important global application it enables people around the globe to interact when distance separates them. In order for the participants in a videoconference call to interact naturally, the end-to-end delay should be below human perception; even though an objective and unique figure cannot be set, 100 ms is widely recognized as the desired one-way delay requirement for interaction. Since the global propagation delay can be about 100 ms, the actual end-to-end delay budget available to the system designer (excluding propagation delay) can be no more than 10 ms. We identify the components of the end-to-end delay in various configurations with the objective of understanding how it can be kept below the desired 10-ms bound. We analyze these components step-by-step through six system configurations obtained by combining three generic network architectures with two video encoding schemes. We study the transmission of raw video and variable bit rate (VBR) MPEG video encoding over 1) circuit switching; 2) synchronous packet switching; and 3) asynchronous packet switching. In addition, we show that constant bit rate (CBR) MPEG encoding delivers unacceptable delay on the order of the group of pictures (GOP) time interval when maximizing quality for static scenes. This study aims at showing that having a global common time reference, together with time-driven priority (TDP) and VBR MPEG video encoding, provides adequate end-to-end delay, which is 1) below 10 ms; 2) independent of the network instant load; and 3) independent of the connection rate. The resulting end-to-end delay (excluding propagation delay) can be smaller than the video frame period, which is better than what can be obtained with circuit switching. Index Terms End-to-end delay, MPEG, performance guarantees, quality of service, time-driven priority, videoconference. I. INTRODUCTION INTERACTIVE real-time applications over packet-switched networks are challenging. Even though in some cases a poor service can be tolerated e.g., if it is charged at a low price the focus of this work is on a high-quality service. One of the key features of such a service is to enable natural interaction that requires the end-to-end delay to be below human perception. Various studies concluded that for natural hearing this delay should be approximately 100 ms [13]. Even though an objective and unique figure does not exist, a 100-ms delay ensures full satisfaction to all users. While a lower end-to-end delay cannot be appreciated, delays above 100 ms will be noticed by some users Manuscript received February 12, 1998; revised February 23, 1999 and December 14, 1999; approved by IEEE/ACM TRANSACTIONS ON NETWORKING Editor M. El Zarki. M. Baldi is with the Department of Computer Science, Politecnico di Torino, Torino 10129, Italy, and with Synchrodyne Networks, Inc., New York, NY USA. Y. Ofek is with Synchrodyne Networks, Inc., New York, NY USA. Publisher Item Identifier S (00) and will lead them to look for a better service; eventually, all users will look for a better service. Thus, service providers that can guarantee a 100-ms end-to-end delay will have a distinct market advantage. The problem that we study in this work is videoconferencing in which voice and video should be synchronized (a.k.a. lip-sync), thus, the end-to-end delay of the video should be below 100 ms as well. Since the global propagation delay can be about 100 ms, the actual end-to-end delay budget available to the system designer (excluding propagation delay) can be no more than 10 ms. The video stream requires high capacity and since network resources are limited, the video pictures should be compressed. Compression can be costly in terms of end-to-end delay. In this work we assumed MPEG encoding [10], since it is one of the most popular compression techniques. Other encoding techniques such as Motion JPEG [7] and H.261 [11] are possible, but they are not in the scope of this paper and are left for further research. Here we show that MPEG constant bit rate (CBR) video encoding is not advisable for high-quality videoconferencing applications. As discussed in Section III-B, efficient compression with the elimination of temporal redundancy introduces an unacceptable delay. At the expense of a higher bit-rate video stream, lower delay is obtained by not eliminating temporal redundancy the MPEG encoder works therefore like a Motion JPEG encoder. Intuitively, for a small end-to-end delay, a picture (video frame) is captured, compressed, and sent once there are enough data units to send them in a packet over the network. This approach can result in a short end-to-end delay. However, it raises some key questions: 1) when will the compressed video data units be ready to be sent (in general, the data units generation during compression is difficult to predict); and 2) how many data units will be generated after the compression of each picture. Both pieces of information are needed in order to reserve communication resources inside the network. In other words, the difficulty arises since the time data units are produced and the amount of data units produced may change from picture to picture. In this paper, we discuss the two problems in details and suggest some solutions which provide deterministic and predictable global quality-of-service (QoS) guarantees. The heart of these solutions is the way in which packet queueing and forwarding is managed by nodes. Some solutions are based on nodes having a common time reference obtained from the global positioning system (GPS) [3] and using it to control packet forwarding; this allows queueing to be reduced in the network nodes. Other approaches are based on asynchronous packet forwarding and require special techniques, /00$ IEEE

2 480 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 8, NO. 4, AUGUST 2000 Fig. 1. delay. Model of a videoconferencing system. P: processing delay. PR: processing resynchronization delay. N: network delay. NR: network resynchronization such as packet generalized processor sharing (PGPS) [17], to bound the packet delay by using scheduling algorithms that approximate the fluid flow service model. Such schemes guarantee a bound on the queueing delay which is inversely proportional to the connection (or session) rate, proportional to the number of nodes traversed by that connection, and proportional to the packet size. The common time reference enables the solution with the lowest end-to-end delay bound and jitter. However, other solutions can be used when the end-to-end delay requirement is not stringent, i.e., when users can tolerate poor interactivity due to a large delay bound, or variable quality due to nondeterministically controlled delay. Designing the network for stringent delay requirements can be beneficial to various system aspects, other than user-perceived quality. One of the main advantages of small delay jitter is small buffers inside the network and at the receiver side. Networks with high speed links require buffers with short access time, which can be expensive, and therefore, small delay jitter will save money. In the future, when virtual reality applications will become real (rather than virtual) the system constraints, such as delay and loss, will be even more rigid. Therefore, in the coming years, the competition among various network vendors will be more than just the capability to provide the service: it will become primarily a competition to provide better quality of network services. Some of the techniques discussed in this manuscript will indeed provide better quality at a lower price (because of less stringent buffering requirements, for example). II. THE MODEL In this section we identify the components of the end-to-end delay of a videoconferencing system for a number of relevant system configurations. In Section III, we analyze each delay component for the various configurations. The model focuses on video rather than audio for two reasons: 1) The audio sampling rate is typically 8 khz, which means that a voice sample is produced every 125 s, while the video sampling rate is much lower, typically 5 30 pictures/s. In general, with higher sampling rate it is possible to obtain shorter delays. Since audio samples are encoded on few bits (at most 8 bits per sample), the main limitation for audio delay is the time needed to obtain enough samples to build up a packet with reasonably low overhead. This delay can be kept short by using packets with a small size header and consequently a small size payload, e.g., the ATM cell payload is only 48 bytes. 2) The audio bandwidth requirements are small relative to (good quality) video, and therefore, audio compression is not necessary in high-speed networks, while video will have to be compressed in the foreseeable future. A. High-Level Delay Components Fig. 1 shows the model of a videoconferencing system. The end-to-end delay of the system is the time elapsed from when a video image is captured by the video camera at the sender side until when it is displayed on the monitor at the receiver (upper arrow in Fig. 1). In order for the participants in the videoconference call to be able to interact naturally we have the following objective: Objective 1: The end-to-end delay (including propagation delay) should be below 100 ms. For delivering high visual quality, video frames should be displayed on the receiver s monitor at the same fixed pace they have been captured. This leads to the second objective (see Fig. 1): Objective 2 Continuous play: The receiver displays pictures (plays audio samples) continuously at the same rate that they have been captured by the sender. This means that the end-to-end delay between capture and display is constant. The end-to-end delay is modeled with four high-level components whose values depend on the system configuration. The segmented arrow in Fig. 1 shows which function in the system introduces each delay component. 1) A processing delay (P) is introduced on both the sender and receiver sides. It may encompass, for example, the time spent in compressing and decompressing of pictures. 2) The network delay (N) is the time needed to move data units from a source to the other videoconference participant(s). The network delay also includes the protocol processing in both the sender and receiver(s). The above delay components can vary during a videoconference call. In order to meet Objective 2 (i.e., constant end-to-end delay) these variations should be smoothed out before pictures are displayed at the receiver. We identify two resynchronization delay components, which are typically realized by some sort of replay buffer. All pictures when exiting this buffer have experienced the same delay from the time when they were captured.

3 BALDI AND OFEK: END-TO-END DELAY ANALYSIS OF VIDEOCONFERENCING OVER PACKET-SWITCHED NETWORKS 481 3) The processing resynchronization delay (PR) cancels the delay variations in generating the compressed video data units. 4) The network resynchronization delay (NR) cancels the variations of the delay experienced in the network (e.g., the delay jitter due to queueing in network nodes). Thus, Objectives 1 and 2 can be summarized using the four delay components in the following way: after resynchronization, the end-to-end delay including propagation delay is B. Analysis Methodology In this work we analyze the end-to-end delay by going through a number of configurations, each adding one or more components to the end-to-end delay. The system configurations differ for the network architecture and the video encoding technique exploited. We consider three network architectures and three video coding techniques. Since we conclude that one of the techniques is not suitable for videoconferencing, we are left with the six system configurations which are studied in Section III. The network architectures are described in Section II-C. The following three video encoding schemes are considered: 1) Raw video: The three color components of each picture are digitally sampled and the resulting data units are transmitted over the network with no delay. In order to reduce the network delay, the bits encoding each frame should be sent as soon as they exit the capture card. Since the capture card provides all the bits within few milliseconds, the traffic generated by the source may be concentrated in this short time interval, i.e., it is bursty. 2) CBR MPEG: Pictures are encoded according to the MPEG standard. Compression is obtained by eliminating spatial (within each picture) and temporal (between subsequent pictures) redundancy. The amount of bits needed to encode each picture (a.k.a. picture size) is not known in advance and is highly variable. Since the picture rate is constant, bits are produced at a variable rate. A buffer is used to smooth the production bit rate: bits exit this buffer (and enter the network) at a constant rate. The encoder is controlled according to the fill level of the buffer in order to prevent it from either overflowing or underflowing. The delay introduced by the buffers cannot be kept below 10 ms for CBR MPEG, as discussed in detail in Section III-B. 3) Variable Bit Rate (VBR) MPEG: Only a small buffer is exploited at the output of the encoder for assembling data units which may exit the encoder shortly after they are produced. The rate of the resulting compressed stream is highly variable. C. Network Architectures Three network architectures are considered. The first is circuit switching, which is a fully synchronous network: routing and flow control are accomplished using time. The second is asynchronous packet switching with no notion of a global common time reference. The third is a combination of the previous two, called time-driven priority (TDP); it has a global common time reference that is used only for flow control and not for routing. 1) Circuit Switching: A fixed amount of link capacity is assigned to each videoconference call by means of time division multiplexing. At the time a data unit (e.g., a byte) is transmitted from its source, it is possible to predict deterministically when it will exit any switch along its route. The time resolution of this advanced knowledge is much shorter than the data unit transmission time. Consequently, network nodes introduce a small delay (a few microseconds). 2) Asynchronous Packet Switching: In packet-switched networks data is gathered in packets which are sent to an ingress switch or router. Switches forward packets toward the destination while statistically multiplexing packets from different sources which are forwarded over the same link. When a packet has to be forwarded on a busy link, it is delayed until the link is available. This delay is called (network) queueing delay. Queueing delay has a high variability since the time spent in an output buffer depends on the packets already in this buffer and in other buffers of the same output port. Thus, the distribution of the queueing delay experienced by packets throughout the videoconference call is determined by the resource allocation policy, the scheduling algorithms used for buffer management, and the overall network traffic characteristics. Queueing delay often accounts for a large portion of the network delay (see Section III-A-3 for details on network delay components) and its high variability gives a major contribution to the network delay variation, a.k.a. network jitter. The maximum jitter is defined, in the context of this work, as the difference between maximum and minimum delay. The network resynchronization delay should be between zero (for packets having experienced maximum network delay) and the maximum jitter (for packets having experienced minimum network delay). Actually, the replay buffer introduces an excess (network) resynchronization delay up to the maximum network jitter (see Section III-A-3 for a detailed explanation). As a result, the network delay possibly contributes with its maximum value plus its maximum jitter. 3) Time-Driven Priority: TDP [14] gives higher priority to real-time traffic in a periodic fashion in order to provide the following properties for real-time traffic: 1) bounded delay, which is independent of the best-effort data traffic; 2) constant bound on the jitter, which is independent of the network size; and 3) either deterministic no-loss or probabilistic control of the loss due to congestion inside the network. For example, for real-time service it is possible to ensure deterministically no (loss due to) congestion inside the network. Moreover, this can be achieved under a full link utilization and without adversely affecting the QoS. TDP is a multiplexing scheme aimed at sharing link capacity while guaranteeing users against uncontrolled delays (or even losses) due to contention in accessing the network s links. The time is divided into time frames (TFs) of fixed duration (a typical choice is = 125 s). Given the link capacity, in each TF a fixed amount of bits can be sent on a link. Assuming small propagation delay, a real-time packet is forwarded one hop every TF. Since packets are not buffered for uncontrolled time, in the more general case with arbitrary propagation

4 482 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 8, NO. 4, AUGUST 2000 Fig. 2. Time-driven priority (TDP) packet forwarding. delay, each packet takes a fixed number of TFs to move from the output buffer of an intermediate node to the output buffer of the following one on the path to the destination, as shown in Fig. 2. This is also called RISC-like forwarding because packets move through the network in the same step-lock fashion as instructions advance in the pipeline of a RISC processor. III. END-TO-END DELAY ANALYSIS This section presents the end-to-end delay of the six system configurations following the delay model presented in Fig. 1. For clarity, two of the delay elements shown in Fig. 1 are not included in the analysis, since they are the same in all configurations. They are briefly described as follows: 1) Capture delay: The frame grabber introduces a constant delay on the order of few milliseconds. 2) Presentation delay: As a picture is ready to be displayed on the receiver side, it is inserted into the video frame buffer which is periodically scanned by the video adaptor to trace the image on the screen. This introduces a presentation delay which can be up to 17 ms, assuming the refreshing frequency to be 60 Hz (i.e., 1/60 = ms). The presentation delay can be eliminated by synchronizing the decoder, the video controller inside the receiver, and the capture card, as shown in Fig. 1. This requires synchronization between network and decoder, receiver network interface and sender network interface, network and encoder, encoder and capture card. This end-to-end synchronization, otherwise very hard to implement, can be easily obtained by using a common time reference, e.g., from the GPS [3]. A. Raw Video The delay analysis of raw video is interesting since it concerns a reduced set of delay components, which are the network (N) and the network resynchronization (NR) delays. There are only two delay components since there is no compression, and therefore, the processing (P) and processing resynchronization (PR) delays are null. 1 1) Circuit Switching: In circuit-switched networks it is assumed that the video transmission is continuously using the bandwidth,, allocated to this circuit. Fig. 3 shows, for each frame, the resulting timing (and the rate) of the production of 1 Analog-to-digital and digital-to-analog conversions of pictures require few milliseconds which is a short time in comparison to the other delay components. Fig. 3. Raw video encoding and transmission over a circuit switched connection. bits by the capture card and of their transmission over the network. This introduces a network shaping delay where is the picture size in bits. The end-to-end delay is constituted by a single component (N depicted in Fig. 1) given by where is the propagation delay, is the circuit switching delay (typically, a few microseconds). The minimum circuit bandwidth required for the transmission of raw video is The end-to-end delay can be reduced by allocating a larger bandwidth to the circuit. In this case, the circuit is busy only for a time in each video frame period. As a result, the remaining time is wasted and no other connection can exploit the reserved resource left unused. If minimum bandwidth is to be allocated for the videoconference call, the end-to-end delay is as large as one video frame period. Therefore, the lower the video frame rate, the larger the end-to-end delay. For example, the minimum bandwidth required to send raw QCIF pictures at 15 frames/s, is 4.5 Mb/s and the resulting shaping delay is 67 ms. However, if more bandwidth is allocated to decrease the network shaping delay to 30 ms, more than 50% of the allocated bandwidth is wasted because the circuit is idle for half of the video frame period. 2) Time-Driven Priority: Raw video can be sent over a packet-switched network with TDP by inserting each picture into one or more packets which are transmitted during a TF. 2 During the TFs between the transmission of two subsequent pictures of the same session, capacity can be reserved to other real-time sessions, as shown in Fig. 4. All the unused capacity (both reserved and unreserved) can be exploited for the transmission of best-effort traffic. 2 If the link capacity is not large enough to allow a picture to be transmitted in a single TF, it is sent over more successive TFs. This introduces a network shaping delay given by the number of TFs needed to transmit each picture. (1)

5 BALDI AND OFEK: END-TO-END DELAY ANALYSIS OF VIDEOCONFERENCING OVER PACKET-SWITCHED NETWORKS 483 Fig. 5. Minimum network resynchronization delay. Fig. 4. Generation of raw pictures and transmission over a network with TDP. The network delay is the only component of the end-to-end delay. It can be expressed as (where is a function of the number of nodes and the processing delay inside each node) plus the propagation delay. The end-to-end delay is In TDP the presentation delay is zero since the frame grabber and the video display adaptor are synchronized. Resource reservation is based on the definition of a time cycle which encompasses a predefined number of TFs: all the nodes share the same knowledge of the ordinal position of the current TF inside the time cycle. 3 Bandwidth is allocated to a sender/receiver pair, by properly reserving (a fraction of) the link capacity during a number of TFs per time cycle on each link on the path from sender to receiver. 4 In order for intermediate nodes to perform the RISC-like forwarding, the TFs on a link must be chosen according to the TFs reserved on the upstream link, and the time needed for a packet to be transmitted from the output buffer of the upstream node to the output buffer of the considered node (see more details in [14]). 3) Asynchronous Packet Switching: A picture, split over one or more packets, is sent through the network from the source to a packet-switching node introducing a transmission delay, where is the picture size and is the link bandwidth or capacity. For example, the transmission of a QCIF image over a T3 link (45 Mb/s) introduces a delay of 6.7 ms. In the network, packets experience a fixed propagation delay and a variable queueing delay; all the delays mentioned so far are part of the network delay component (N). Due to the real-time requirements of the video stream, a replay buffer is needed at the receiver to compensate for the variation in the queueing delay. The compensation is obtained by delaying the samples that have experienced a queueing delay shorter than the maximum ; the delay introduced is part of the network resynchronization delay. As a result of the compensation, the sum of the queueing delay and its compensation experienced by each sample is the maximum queueing delay. 3 This can be easily implemented with GPS [3]. 4 The TFs during which capacity is allocated to a videoconference call are said to be reserved to the call. (2) Actually, the compensation of the queueing delay introduces also the excess resynchronization delay constant over the duration of the whole conference; is the maximum variation of the queueing delay, i.e., the difference between the maximum and the minimum queueing delay. The excess resynchronization delay is introduced because when a packet is received the actual delay it experienced in the network is not known, as explained in detail in [1]. The compensation is implemented by delaying the first packet of a flow by and then retrieving from the replay buffer the subsequent packets at a constant rate. Fig. 5 shows the resynchronization of a packet stream, the first packet of which has experienced minimum queueing delay : the upper diagram shows the arrival time of pictures to the replay buffer, while the lower shows the exit time. If the network interfaces of sender and receiver are not synchronized, the latter is not able to determine the queueing delay experienced by a packet. In particular, not knowing the delay experienced by the first packet received for the videoconference call, the receiver buffers it for a time that allows resynchronization in the worst case, i.e., it is assumed that the packet has experienced the minimum queueing delay and is buffered for a time The following packets are resynchronized accordingly, as shown in Fig. 5, because they are retrieved from the replay buffer at the constant pace at which pictures are displayed. The upper diagram of Fig. 6 shows the arrival time of pictures when the first packet experiences an actual queueing delay. The middle diagram shows the timing of packets exiting; the receiver, not knowing the actual queueing delay experienced by the first packet, delays it by according to (3). As a consequence, the overall delay experienced by the following packets due to queueing and resynchronization is between and. If the receiver knows the actual delay experienced by the first packet (e.g., sender and receiver have a common time reference and the packet contains a time stamp indicating when it was sent), a packet that has already experienced maximum queueing delay in the network is not further delayed in the replay buffer and the exit times are those depicted in the lower diagram in Fig. 6. Thus, if sender and receiver do not share a common time reference, the end-to-end delay experienced by he ser of the (3)

6 484 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 8, NO. 4, AUGUST 2000 Fig. 6. Real network resynchronization delay. videoconferencing system is The compensation delay introduced by the replay buffer and the related error constitute the network resynchronization delay component (NR identified in Fig. 1), while the other terms are part of the network delay component N. It can be worth pointing out that the contribution accounts partly for the N component (since it contains the queueing delay) and partly for the NR component (since it contains the compensation delay). The excess resynchronization delay can be eliminated if the sender and the receiver have a common time reference. The maximum queueing delay is usually much larger than both all other network delay components and the minimum queueing delay (that is given by the sum of the transmission delay over the traversed links). Consequently, and, due to jitter compensation, the maximum queueing delay can contribute twice (namely, by itself and as excess resynchronization delay) to the end-to-end delay [1]. Schemes like PGPS [17] are being proposed for preventing loss and bounding the queueing delay. Such schemes provide a bound that is inversely proportional to the bandwidth allocated to the session, and proportional to the packet size and the number of hops. This is stated, for example, by the general result on PGPS Networks in [17, Section X, p. 146]. In particular, [17, Eqs. 37, 38, and 39, p. 148], which have the following general structure: PGPS-Delay-bound-connection, where is the packet size, is the number of hops and is the rate of connection. Note that some of the results reported in [17] are based on several timing assumptions, such as, that the delay between nodes is zero or constant. Such an assumption will require the synchronization of the local clocks of all the nodes, which is equivalent to the global common time reference used for TDP. The contribution to the end-to-end delay due to the queueing delay and its compensation can be reduced by underdimensioning the replay buffer, thus having the queueing delay and the excess resynchronization delay contributing with a value. As a consequence, all the (parts of) pictures experiencing a network delay larger than are discarded at the receiver at the expenses of the visual quality of the received video stream. is some percentile of the queueing delay chosen as to guarantee that the percentage of discarded packets does not affect visual quality. Some videoconferencing applications explicitly designed for operation over asynchronous packet-switched networks adapt the resynchronization delay introduced by the replay buffer to the instantaneous distribution of the network delay experienced by packets [24]. This results in variable visual quality and user perceived delays, and does not comply with Objective 2. Losses in network nodes due to congestion and buffer overflow also decrease the perceived quality of a videoconference call. In order to reduce buffer overflows and control the distribution of queueing delay over the duration of videoconference calls, resources are reserved in the nodes along the connection path and access to the network is controlled. This limits the amount of guaranteed traffic routable over the same link. Since transmission of large amounts of data at link speed (bursts) makes queues grow suddenly, bursty sources require a large amount of resources to be allocated and significantly reduce the overall amount of real-time traffic the network can support. Source burstiness can be reduced through traffic shaping at the network boundaries. Traffic-shaping mechanisms like, for example, the leaky bucket [2], guarantee an average bandwidth to the source while keeping the burstiness below a predefined value. 5 This introduces a shaping delay where is the largest burst size, i.e., the maximum number of bits which can be sent at the full link speed. On one hand, the traffic shaping at the boundary of the network reduces the buffer requirements in the nodes, the queueing delay in the network and its variability (i.e., both the N and NR components of the end-to-end delay model proposed in Fig. 1); on the other hand, it introduces a variable shaping delay that is compensated on the receiver side (i.e., it contributes to both the N and NR components of the end-to-end delay model proposed in Fig. 1). In summary, the end-to-end delay can be expressed as where is the size of packets sent into the network. B. Why VBR MPEG? Compression, although it reduces the transmission delay, may have large processing and processing resynchronization delays. Moreover, if the naturally variable rate of a highly compressed stream is to be converted into a constant one, a significant contribution to the processing resynchronization delay must be further introduced, thus adversely affecting the end-to-end delay. The objective of the following discussion is to present and justify the rationale for recommending the use of VBR MPEG, rather than CBR MPEG, for videoconferencing applications. The motivation is that VBR encoding can provide high compression while 5 If a leaky bucket is exploited, B is its token generation rate and A is the token pool size. (4)

7 BALDI AND OFEK: END-TO-END DELAY ANALYSIS OF VIDEOCONFERENCING OVER PACKET-SWITCHED NETWORKS 485 keeping very short the contribution to the end-to-end delay due to the processing resynchronization component. 6 1) MPEG Overview: The Motion Picture Expert Group (MPEG) [10], [8] video encoding standard was designed for digital storage of quality video for later replaying. The encoder receives a sequence of digitalized pictures and encodes each of them in one of two different ways. 7 Intraframe Coding eliminates spatial redundancy inside pictures. The picture (the luminance and the two chrominance components) is divided in 8x8 pixel blocks. On each block a discrete cosine transform (DCT) is performed and an 8 8 matrix of coefficients is devised. Each coefficient is quantized by integer dividing it by an integer called quantization step size. The result of the quantization is run-length encoded to gain further compression thanks to the many zero-valued coefficients. Finally, the run-length encoded symbol sequence is Huffman encoded. The resulting encoded picture is called an I-frame. Predictive Coding eliminates temporal redundancy between a picture and the previous one. The picture is divided into macro blocks (MBs), each composed of a pixel matrix of luminance information and the two corresponding 8 8 pixel blocks of the two chrominance components. Motion estimation is performed for each MB, i.e., the previous picture is searched for a similar MB. If this MB is found, the pixel by pixel difference between the actual MB and the reference one is calculated and coded by performing DCT, quantization, run-length encoding, and Huffman encoding. If a similar MB is not found, each block of the MB is encoded like a block in an I-frame. The obtained encoded picture is called a P-frame and it is typically 2 to 4 times smaller than an I-frame. The more similar two subsequent pictures are, the higher the probability of finding in the reference picture a MB similar to the one being coded. Subsequent pictures are similar if the scene is slow moving, thus not changing much from a video frame period to the other. In summary, predictive coding delivers more compression on slow scenes. A video sequence may be compressed by encoding one picture out of as an I-frame, and the remaining pictures as P-frames; the sequence of pictures is called a group of pictures (GOP). The larger, the smaller the amount of bits needed to encode each video sequence that contains a good deal of temporal redundancy. If the network introduces an error in an encoded I-frame, the error may propagate into the entire GOP. The next I-frame is the first picture not to be affected by such error. Thus, the larger the more damage an error can cause. 2) Controlling the Stream Rate: Due to the difference between I-frames and P-frames, the rate of the bit stream produced by the encoder has high variability. Fig. 7 shows the amount of bits produced by the software MPEG encoder dvdenc [16] during the encoding of the Cheerleaders video sequence. It shows a group of cheerleaders in a stadium, being therefore a scene with a lot of motion on a background with many details. 6 A detailed discussion can be found in [1]. 7 Actually, a third type of encoding, called bidirectional predictive coding, exists. Before a picture can be coded, a reference subsequent picture must be captured and coded. This introduces a delay of some frame periods that we deem not acceptable given the 100-ms end-to-end delay bound. Thus, this type of compression is not considered here. Fig. 7. Fig. 8. Natural MPEG encoding of the Cheerleaders sequence. High level model of a CBR MPEG encoder. A CBR MPEG stream is obtained by filling a buffer with the output of the basic encoding process and retrieving bits at a constant target rate, as shown in Fig. 8. This buffering process introduces a sensible variability of the processing delay component of each picture which must be compensated in the decoder thus introducing a large processing resynchronization delay component. A rate control function (in principle) monitors the fullness of the encoder buffer and adjusts the natural bit rate in order to prevent it from underflowing or overflowing. The rate control function tunes the bit production of the encoder to grant the stream compliance according to a model of a system target decoder buffer (see Fig. 8) whose dimension is included in the MPEG stream [10]. The bit production can be modified acting on the step size of a quantizer used at the last stage of the encoding process. 8 3) Delay and Picture Dimension: The CBR encoder contributes to the processing delay with, being the size of an encoded picture. Since picture size is not constant, this contribution is variable and must be compensated at the receiver. The resulting overall contribution (after resynchronization) to the end-to-end delay is called coding shaping delay. where is the maximum picture size over the whole sequence. In order for the video stream to be continuous (actually having a constant rate), ; otherwise there would be a time interval between two subsequent frames during which no bits exit the encoder. As a consequence, CBR MPEG encoding 8 See [8] for further information on the MPEG encoding process.

8 486 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 8, NO. 4, AUGUST 2000 Fig. 9. CBR MPEG encoding with N =15. always introduces a delay larger than the video frame period. With reference to the model depicted in Fig. 1, contributes to the P and PR components. Fig. 9 shows the picture size obtained when encoding four video sequences with the dvdenc [16] CBR MPEG encoder. Each sequence is encoded at three different target rates. The maximum picture size in each sequence determines the lower bound on the coding shaping delay. In order to keep constant the rate of the video stream (i.e., avoid the encoder buffer underflow), the encoder must either limit the compression and keep P-frames from becoming too small, or increase the size of I-frames as the P-frames become smaller. P-frame dimension is determined by the chosen encoding parameters and by the motion in the scene: the slower the scene, the larger the amount of temporal redundancy, the smaller the size of P-frames if such redundancy is eliminated. Videoconferencing scenes are usually quite static: P-frames are expected to be small and I-frames consequently large. In order to obtain a confirmation from experimental data, we encoded two completely static scenes built by replicating 120 times the same picture of the Cheerleaders and Hockey sequences, respectively. Fig. 10 shows the resulting picture size for the same target rates used in the previous experiment, deploying the encoding parameters of dvdenc. As expected, using these encoding parameters, the maximum picture size, i.e., the lower bound on the coding shaping delay, has significantly increased with respect to Fig. 9, especially at low bit rates. The maximum picture size throughout a sequence is not known in advance. Nevertheless, when dealing with real-time video (like in a videoconferencing scenario) the encoding delay bound must be known before starting the videoconference call in order to meet Objective 2 (continuous playing) by introducing the proper resynchronization delay. The following of this section is devoted to identifying the maximum picture size given the characteristics of the CBR MPEG encoder. We show that the corresponding coding shaping delay when applying CBR MPEG to videoconferencing is unacceptable for meeting Objective 1 (end-to-end delay less than 100 ms). 4) Delay on the Order of the GOP Duration: Since the rate control function aims at avoiding the encoder buffer overflow, the picture size is upper bounded by the buffer dimension. It follows that the coding shaping delay can be reduced by exploiting a small encoder buffer. Hence, the issue becomes how to reasonably dimension the encoder buffer. Whenever a picture smaller than is produced, the buffer must contain a backlog large enough to guarantee the continuity of the stream. Thus, the smaller the buffer, the less picture size can vary. In principle, in order to deliver maximum quality the buffer should be large enough to allow P-frames to be encoded with no bits when the image is completely static. This is particularly important when dealing with videoconferencing because the camera can be pointed, for example, over a blackboard thus capturing a completely static scene. The fewer bits used to en-

9 BALDI AND OFEK: END-TO-END DELAY ANALYSIS OF VIDEOCONFERENCING OVER PACKET-SWITCHED NETWORKS 487 Fig. 10. CBR MPEG encoding of a static scene. Fig. 11. Bit production of a noncontrolled VBR MPEG encoder Processed bits. Encoder output. code P-frames, the more can be dedicated to I-frame (thus creating the backlog in the buffer) and the higher the quality of the resulting image. Quality of static images is particularly critical since human eye is more sensitive to errors on static images, than on moving scenes. Null-size P-frames can be produced if an I-frame can be as large as the whole amount of bits sent during a GOP. That is, in order to deliver maximum quality of static images, the encoder buffer size must be at least the GOP size; this results in a coding shaping delay on the order of the GOP duration. Such a coding shaping delay is not acceptable when aiming at Objective 1: for example, if operating at 30 frames/s with a GOP size of 15 pictures, the coding shaping delay is 500 ms. An encoder designed to keep the introduced delay bounded may avoid eliminating temporal redundancy in P-frames, by encoding many macro blocks as in I-frames. In this way the size of I-frames is smaller and the coding shaping delay limited consequently. However, the quality of a static scene, like the one resulting from pointing the camera over a blackboard, is not maximized. In summary, CBR MPEG encoding may not be the most advisable scheme for high-quality videoconferencing; VBR MPEG is evaluated in the following and it seems to provide an appealing solution. C. Transmission of VBR MPEG While the CBR encoder introduces an unacceptable delay in the encoder, a VBR video stream may impact the network performance leading to either high delay, or high loss, or the need for an overallocation of communication resources in some of the configurations. The rest of this section examines the resources required for the transmission of VBR MPEG video over the three network architectures, and the end-to-end delay of the resulting videoconferencing system. 1) Circuit Switching: The MPEG encoder produces bits with a timing similar to that shown in Fig. 11 as it encodes I-frames and P-frames. The time required to encode a picture and the amount of bits produced are not constant. In principle, if the videoconference call is allocated a circuit with bandwidth larger than the maximum instantaneous rate of the encoder, bits are transmitted as soon as they are produced and bits get to the decoder at the same rate they had been produced after having experienced the constant propagation and switching delays. The time needed to encode a picture is not constant; the decoder introduces a processing resynchronization delay to keep constant the time between decoded pictures. As a result, each picture experiences an overall coding-decoding-resynchronization delay which is the maximum time required to encode and decode a picture. The end-to-end delay is given by where the first term is the overall coding-decoding-resynchronization delay and contributes to the processing delay P, 9 while the other two terms contribute to the N component. Since we are dealing with real-time video, a picture should be encoded (decoded) within the video frame period, i.e., any encoder will feature. Thus, if a scene is captured at 30 frames/s and the end-to-end delay objective is Actually, CD contributes also to the PR component, but since the variability of the P component is relatively small, the contribution of PR can be neglected. See [1] for details.

10 488 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 8, NO. 4, AUGUST 2000 Fig. 12. Transmission of VBR MPEG video with TDP. ms, the delay budget available for propagation is 40 ms which corresponds to a span of 8,000 Km (like a transoceanic call). A faster encoder features a lower and consequently enables a larger span. The system configuration discussed in this section provides a lower bound on the end-to-end delay in a videoconferencing system exploiting MPEG compression. Nevertheless, it is not practical since the bandwidth of the circuit allocated to the videoconference call is only partially used and the unused fraction is wasted. Moreover, since the bandwidth of the circuit is equal to the peak rate of the encoder, encoding P-frames at a lower bit rate does not provide any advantage. Since motion estimation is the most time consuming function of the encoding process, is significantly reduced by exploiting only intraframe coding. This corresponds to deploying a Motion JPEG encoder. Reducing bandwidth occupation introduces a larger delay since transmission over a circuit with a bandwidth smaller than the peak rate of the encoder introduces a network shaping delay. 2) Time-Driven Priority: As soon as all the bits for encoding a picture are produced by the encoder, they are inserted into a packet and sent at the full speed of the ingress link, as depicted in Fig. 12. The end-to-end delay of the system is given by where is the number of TFs a packet takes to travel from sender to receiver. With reference to the model of the videoconferencing system depicted in Fig. 1, the first term contributes to the P component (and in a negligible way to PR, while the second term contributes to the N component. Equation (5) is the actual end-to-end delay only if the nodes on the path from sender to receiver perform RISC-like forwarding of packets. To guarantee the fixed network delay and loss-free delivery, resources must be allocated in the network and video frames sent during reserved TFs. To reserve resources in packet-switched networks with TDP, the amount of data to be sent and their timing must be known at reservation time so that the proper fraction of link capacity can be reserved during the proper TFs. The amount of bits reserved should be larger than (and as close as possible to) the dimension of the (5) picture being sent. As it was shown in Fig. 7, picture dimension is not known in advance and resource reservation may not be accurate. If during a TF a user sends more bits than the reserved amount, the network does not provide any guarantee on the delivery of the excess data units. If, on the other hand, the videoconferencing application uses only a fraction of the reserved capacity, the leftover bandwidth can be used by best-effort traffic and is not wasted (unlike circuit switching). Even though this is acceptable from the network point of view, the solution is not optimal for the user who is possibly paying for the allocated bandwidth and would like to use it all by himself. The rest of the section concerns: 1) determining the amount of bits to be reserved for both types of pictures given the bandwidth to be allocated to a videoconference call; 2) tuning the encoding process in order to control picture dimension so that the videoconferencing system never uses more bandwidth than the allocated amount and exploits as much of it as possible; 3) the impact of scheduling (i.e., the choice of the TFs to be reserved) on the end-to-end delay. Even though some configurations can deliver unacceptable end-to-end delay, if the system is adequately equipped and operated, its performance is actually given by (5). 1) Choosing a Bound on Picture Dimension. As explained in Section III-B on MPEG, the slower the motion in the scene being encoded, the larger the dimension of I-frames with respect to P-frames. Since videoconferences are expected to be slow-moving scenes, we propose to reserve different amounts of bits for transmission of I-frames and P-frames. These amounts determine the bandwidth reserved to the videoconference call as where and are the amount of bits reserved for I-frames and P-frames, respectively, is the number of pictures per GOP, and is the GOP duration. As discussed, the relative dimension of I-frames and P-frames yielded by a noncontrolled MPEG encoder depends on the amount of motion in the scene. The picture ratio must then be chosen wisely depending to the amount of motion expected in the scene to be encoded and transmitted. This is, in general, a difficult task, but in the particular case of videoconferencing, scenes are likely to be slow and consequently large. Combining (6) and (7), the amount of bits to be reserved to each frame can be expressed as a function of the bandwidth to be allocated as (6) (7) (8)

11 BALDI AND OFEK: END-TO-END DELAY ANALYSIS OF VIDEOCONFERENCING OVER PACKET-SWITCHED NETWORKS 489 Fig. 13. MPEG encoder controlling frame dimension. 2) Controlling Picture Dimension. The reserved bandwidth is used more efficiently if the encoding process is controlled and picture size is kept below (and as close as possible to) the amount of bits reserved. This amount is an upper bound on picture dimension and can be used to devise a target dimension for each type of video frame. There exist proposals to predict the amount of bits an encoding process is going to produce based on the raw scene [20] or on the portion of stream already produced [12]. However, to get service guarantees from the TDP network, the upper bound must never be exceeded and an adaptive approach based on feedback from the output of the encoder should be exploited. The encoder is extended with a rate control function, as shown in Fig. 13. It tunes the parameters of the basic MPEG coding process so that the dimension of each picture is best fitted to the target size associated with its type. Among the parameters of the MPEG encoding process the quantization step size is the most suited to this purpose. Varying the granularity of quantization throughout the picture delivers nonuniform visual quality. Many approaches have been proposed in the literature for uniformly choosing the quantization parameter [9] possibly taking into account the characteristics of the human visual system [23]. Other authors propose iterative approaches [25], [15] to determine a suitable value of the quantization step size to be used throughout a whole picture. This can lead to coding times not acceptable for real-time encoding as required in videoconferencing applications. [5] proposes to exploit a rate-quantization model to choose a quantization step size for a whole picture. The model is tuned according to the characteristics of the encoded stream already produced. The rate-control algorithm also proposes how to requantize the picture if the yielded dimension is not compliant with the target. The above mentioned approaches have been proposed and analyzed in scenarios different from ours. Thus, we have performed some experiments to prove that picture dimension can be controlled as the proposed network technology requires. The software encoder dvdenc has been augmented with a rate control function which calculates the quantization step size on a MB-by-MB basis given a target picture dimension and a tolerance on it (in terms of maximum and minimum acceptable dimensions). The quantization step size is determined as a function of the amount of bits produced so far and the number of bits expected according to the target Fig. 14. TDP and complex scheduling.. The prototype rate control function has proven to be able to always keep picture size below the given bound or As a consequence, bandwidth can be efficiently allocated by reserving fixed amount of bits for the transmission of pictures of the same type. 3) Complex Scheduling. Scheduling, i.e., the choice of the TFs to be reserved to a videoconference call, is simplified by considering the nature of the application-generating traffic. Different amounts of bits should be reserved in the TFs intended for sending I-frames and those for P-frames. The time cycle must be set to an integer multiple of the GOP period and TFs must be reserved within the time cycle. The choice of the TFs to be reserved on each link on the path between sender and receiver is called complex scheduling. 10 The choice of the TFs impacts both network performance (in terms of maximum number of real-time connections concurrently supported) and the end-to-end delay of the video-conference call. Fig. 14 shows a sample reservation with and. The upper diagram depicts the frame-grabbing time and the lower one shows the amount of bits reserved in the TFs: in one TF and in the following TFs. The capacity for sending the encoded video frames is reserved in the first TF beginning after (the maximum coding delay) from the capture of each picture. This is the optimal schedule which leads to minimum delay as given by (5). In order for the encoded pictures to be ready before the reserved TF, the capture card (as well as the encoder) must be synchronized with the network interface. Lack of synchronization would introduce a variable delay whose maximum value would be the GOP period (namely, ). Also, if the optimal schedule is not feasible, a scheduling shaping delay is introduced which contributes, together with, to the N component identified in the model depicted in Fig. 1. Assuming synchronization between capture card and network interface, the general equation for the end-to-end delay is where is determined when performing the complex scheduling (i.e., when the videoconference call is placed); it can be small if a wise scheduling is performed. 10 This scheduling is said to be complex because the allocated capacity is not the same during all the reserved TFs. (9)

12 490 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 8, NO. 4, AUGUST 2000 Fig. 15. Encoding and complex scheduling. Reducing Decoding Time: Decoding time can be reduced if the decoder does not wait for the whole picture to be received before starting processing it. This requires encoded pictures to be inserted in smaller packets that are sent as soon as they have been assembled. This does not affect any component of the end-to-end delay given by (9) other than. It is thus better to allocate more than one TF per picture and send a packet in each of them (Fig. 15), than to allocate a single TF during which a whole encoded picture is sent (Fig. 14). Choosing these TFs encompasses two nontrivial issues: determining their position inside the time cycle and their number so that the end-to-end delay can be minimized. In the following, these problems are described and the trade-offs between different choices are outlined, but a solution is not proposed because it is out of the scope of this work. When a TF reserved on the sender link begins, enough bits must have been produced by the encoder, otherwise the amount of transmitted data is smaller than the one reserved. This is called encoder underflow and it is critical not just because the allocated bandwidth is underutilized (it can be exploited by besteffort traffic). If the total amount of bits actually produced to encode the picture is very close to the global amount of bits reserved for the picture (over a number of TFs), the remaining reserved TFs will not have enough capacity to carry the other bits that should be sent in the present TF. On the other hand, if the TFs reserved for a picture are chosen later (with respect to the beginning of the video frame period) in order to minimize the risk of encoder underflow, the benefit of using more TFs to transmit a picture is reduced, i.e., the end-to-end delay is increased. 3) Asynchronous Packet Switching: As soon as the encoder produces enough bits to assemble a packet of dimension, the packet is assumed to be sent into the network where it experiences a variable queueing delay. The receiver must exploit a replay buffer to compensate the queueing delay variation, thus introducing the NR component. The end-to-end delay is thus given by where is the maximum time required to encode and decode a picture (contributing to P and to a small PR, is some percentile of the maximum queueing delay, and is the excess resynchronization delay introduced by the replay buffer (contributing to NR). Using small packets and a decoder which starts decoding video frames as soon as data are received can reduce. The end-to-end delay is dominated by that, with reference to the videoconferencing system model depicted in Fig. 1, contributes to the N component and to the NR component. Traffic Shaping at Network Boundaries: Resources are reserved in the network in order to bound the queueing delay (i.e., to reduce N and NR). As discussed in Section III-A-3, resource reservation is more efficient if traffic shaping is performed at the boundaries of the network, even though it introduces a network shaping delay (i.e., it increases N and NR). Resources are reserved in the network based on the traffic description, given in terms of burstiness and average rate, corresponding to the shaped traffic. The network guarantees the quality of the service (i.e., the deterministic bound on the queueing delay or the statistical one ), only if the actual traffic is compliant with the description given at resource reservation time. The delay globally experienced by a picture due to the traffic shaper depends on the natural bit generation rate of the encoder, the implementation of the traffic shaper, and the characteristics of the shaped traffic. The receiver has to compensate it by means of the network resynchronization delay introduced by the replay buffer. Thus, each packet experiences a network shaping delay partly in the traffic shaper and partly in the replay buffer. Equation (4) gives the delay introduced when dealing with raw video. Since devising an analogous equation for VBR MPEGencoded video is a harder task and it is not the goal of this work, we do not analyze in more detail. The end-to-end delay of the videoconferencing system is given by and contributes to the P, N, and NR components of the end-to-end delay model presented in Fig. 1. (A relatively small contribution to the PR component is also present.) Adapting the Encoded Video Stream to the Network: Since the traffic pattern generated by a natural VBR MPEG encoder is not known in advance, it can be incompatible with the shaped traffic description. The noncompliant packets can be discarded by either the traffic shaper itself, or a traffic policing function inside the network [19]. For example, if a leaky bucket is exploited to shape the traffic, the token generation rate and token pool size determine the average rate and the burstiness of the shaped traffic. If the characteristics of the encoded video are not compatible with the values chosen for and the excess traffic must be either discarded or sent in the network as best-effort traffic. Even though a buffer is inserted before the leaky bucket to adapt the video stream to the traffic description, it can overflow if the two are too different. The loss of packets is not acceptable in the transmission of MPEG-encoded video, especially when the GOP is large. Even though techniques have been proposed to limit the effect of loss [6], it should be better for the videoconferencing system to avoid loss in order to deliver the highest possible quality. The MPEG encoding process can be controlled to avoid that the traffic shaper discards or sends as best-effort traffic packets

13 BALDI AND OFEK: END-TO-END DELAY ANALYSIS OF VIDEOCONFERENCING OVER PACKET-SWITCHED NETWORKS 491 TABLE I END-TO-END DELAY FOR THE SYSTEM CONFIGURATIONS CONSIDERED IN THIS WORK Fig. 16. MPEG encoder controlled using feedback from a traffic shaper. that cannot be adapted to the traffic description. A rate control function tunes the parameters of the basic MPEG encoder according to the traffic description used to drive the traffic shaper [22], [21]. Due to the unpredictable output of MPEG encoders, this approach does not guarantee against packets not compliant with the traffic description. Alternatively, a rate-control function can tune the MPEG encoder parameters based on feedback information received from the traffic shaper (e.g., the fullness of the buffer preceding a leaky bucket) [18], [4], as shown in Fig. 16. If this will significantly degrade the visual quality of pictures, the resource allocation can be renegotiated according to a rate-quantization model that is tuned as the encoding progresses. IV. DISCUSSION In this work we analyze the end-to-end delay of videoconferencing in six system configurations obtained by combining three network technologies with two encoding schemes. The results are summarized in Table I. We study the transmission of raw video and variable bit rate MPEG video over: 1) circuit switching; 2) TDP packet switching; and 3) asynchronous packet switching. In addition, we show that in some encoding configurations, CBR MPEG encoding delivers long delay, which is on the order of the GOP duration. Thus, if the sampling rate is of 30 frames/s and a GOP of 15 pictures is used (i.e., each I-frame is followed by 14 P-frames), the resulting delay is on the order of 500 ms. Given the long distance and high bit-rate requirements of videoconferencing, video compression should be used. Since CBR encoding has unacceptable delay, VBR encoding should be used. In the case of VBR encoding, circuit switching is not practical since the network utilization is very low. Thus, packet switching should be used. However, relying on asynchronous packet switching with statistical multiplexing and first-comefirst-serve queueing discipline can result in high loss and delay jitter under high load conditions. Other queueing disciplines, such as weighted fair queueing, can only guarantee deterministic no loss to CBR traffic, which as mentioned would result in an unacceptably large delay bound. This study shows that having a global common time reference can be used for implementing TDP with complex periodicity scheduling for transporting VBR MPEG encoding. This will provide adequate deterministic delay bounds which are independent of the network load and the connection rate. In such a system configuration, the end-to-end delay (excluding propagation delay) can be smaller than the video frame period. Furthermore, TDP with complex periodicity scheduling can deterministically ensure no loss (due to congestion) of VBR traffic. These unique results cannot be obtained with circuit switching or any other known alternative schemes. ACKNOWLEDGMENT The authors thank P. Tiwari for providing them with the software MPEG encoder dvdenc and P. Westerink for his kind and useful help in understanding, modifying, and operating the encoder. REFERENCES [1] M. Baldi and Y. Ofek, End-to-end delay of video-conferencing over packet-switched networks, IBM, Res. Rep. RC (91480), Dec [2] M. Butto, E. Cavallero, and A. Tonietti, Effectiveness of the leaky bucket policing mechanism in ATM networks, IEEE J. Select. Areas Commun., vol. 9, pp , Apr [3] P. H. Dana. Global positioning system overview. [Online]. Available: [4] W. Ding and B. Liu, Joint encoder and channel rate control of VBR video over ATM networks, Visual Commun. Image Processing, vol. 2666, pp , 1996.

14 492 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 8, NO. 4, AUGUST 2000 [5], Rate control of MPEG video coding and recording by rate-quantization modeling, IEEE Trans. Circuits Syst. Video Technol., vol. 6, pp , Feb [6] J. Feng, K.-T. Lo, H. Mehrpour, and A. E. Karbowiak, Cell loss concealment method for MPEG video in ATM networks, in GLOBECOM 95, pp [7] B. Fufht, A survey of multimedia compression techniques and standards. Part I. JPEG standard, Real Time Imaging, vol. 1, pp , [8] D. Le Gall, MPEG: A video compression standard for multimedia applications, Commun. ACM, vol. 34, no. 4, pp , Apr [9] C. Horne and A. Puri, Video coding with adaptive quantization and rate control, Visual Commun. Image Processing, vol. 1818, pp , [10] ISO/IEC, Information Technology Coding of Moving Pictures and Associated Audio for Digital Storage Media up to about 1.5 Mbit/s, International Organization for Standardization, [11] ITU-T, Recommendation H.261, Sept [12] S. Jung and J. S. Meditch, Adaptive prediction and smoothing of MPEG video in ATM networks, in IEEE Int. Conf. Communications, 1995, pp [13] G. Karlsson, Asynchronous transfer of video, IEEE Commun. Mag., pp , Aug [14] C.-S. Li, Y. Ofek, A. Segall, and K. Sohraby, Pseudoisochronous cell switching in ATM networks, Comput. Netw. ISDN Syst., vol. 30, [15] L.-J. Lin, A. Ortega, and C.-C. J. Kuo, A gradient based rate control algorithm with applications to MPEG video, 2nd Int. Conf. Image Processiing (ICIP 95), pp , [16] E. Linzer, A robust MPEG-2 rate control algorithm, IBM T. J. Watson Research Center, Yorktown Heights, NY, Tech. Rep., Dept. 924A, unpublished. [17] A. K. Parekh and R. G. Gallager, A generalized processor sharing approach to flow control in integrated services networks: The multiple node case, IEEE/ACM Trans. Networking, vol. 2, pp , Apr [18] D. Reininger, G. Ramamurthy, and D. Raychaudhuri, VBR MPEG video coding with dynamic bandwidth renegotiation, in IEEE Int. Conf. Communications, June 1995, pp [19] I. E. G. Richardson and M. J. Riley, Usage parameter control cell loss effects MPEG video, in IEEE Int. Conf. Communications, June 1995, pp [20] R. M. Rodriguez-Dagnino, M. R. K. Khansari, and A. Leno-Garcia, Prediction of bit rate sequances of encoded video signals, IEEE J. Select. Areas Commun., vol. 9, pp , Apr [21] M. Simon, P. Villegas, J. Caballero, and M. Roser, A general approach to output rate control in video coding, Visual Commun. Image Processing, vol. 1903, pp , [22] S. Singh and S.-S. Chan, A multi-level approach to the transport of MPEG-coded video over ATM and some experiments, in IEEE GLOBECOM 95, 1995, pp [23] A. Sultan and H. A. Latchman, Adaptive quantization scheme for MPEG video coders based on HVS (Human Visual Systems), Visual Commun. Image Processing, vol. 2668, pp , [24] T. Turletti and C. Huitema, Videoconferencing on the Internet, IEEE/ACM Trans. Networking, vol. 4, pp , June [25] L. Wang, Rate control for MPEG video coding, Visual Commun. Image Processing, vol. 2501, pp , Mario Baldi (M 99) received the M.Sc. degree summa cum laude in electrical engineering and the Ph.D. degree in computer and system engineering from Politecnico di Torino, Italy, in 1993 and 1998, respectively. He is Vice President for Protocol Architectures at Synchrodyne Networks, Inc., New York, NY, since October, He is an Assistant Professor, currently on leave of absence, at the Computer Science Department, Politecnico di Torino, since November, He has been a Visiting Researcher at IBM T. J. Watson Research Center, Yorktown Heights, NY, at Columbia University, New York, NY, at International Computer Science Institute (ICSI), Berkeley, CA, and at Synchrodyne, Inc. His research activity has been focused on internetworking techniques and real-time services over packet-switched networks. Yoram Ofek (S 86 M 87) received the B.Sc. degree in electrical engineering from the Technion-Israel Institute of Technology, Israel, in 1979, and the M.Sc. and Ph.D. degrees in electrical engineering from the University of Illinois, Urbana, in 1985 and 1987, respectively. Currently President and CTO of Synchrodyne Networks, Inc., from 1979 to 1982 he was a Research Engineer at RAFAEL, Haifa, Israel, and from 1987 to 1998 he was a Research Scientist at IBM T. J. Watson Research Center, Yorktown Heights, NY. He has initiated, invented, and managed the research activities of five novel network architectures: 1) ring networks with spatial bandwidth reuse with a family of fairness algorithms, MetaRing, which is used as the underlying network for SSA (Serial Storage Architecture ANSI Standard X3T10) and several IBM products; 2) optical hypergraph for combining multiple passive optical stars with a novel conservative code for bit synchronization and global clock synchronization; 3) embedding of virtual rings in arbitrary topology network, the MetaNet for bursty data traffic with no packet loss; 4) global packet networks for real-time and multimedia, which utilize GPS-based synchronization for providing deterministic quality of service (QoS) guarantees; 5) fractional lambda (wavelength) switching for WDM networks, which is the focus of his current work. Dr. Ofek was awarded the IBM Outstanding Innovation Award for his invention of the MetaRing and his contributions to the SSA products.

Time. Objective 1: end to end delay below 100 ms. Shaping & Packetization. Dec. Enc. Network Adaptation. Frame Grabber. Video Adaptor.

Time. Objective 1: end to end delay below 100 ms. Shaping & Packetization. Dec. Enc. Network Adaptation. Frame Grabber. Video Adaptor. End-to-end Delay of Videoconferencing over Packet Switched Networks Mario Baldi Politecnico di Torino, Dipartimento di Automatica e Informatica Yoram Ofek IBM, T. J. Watson Research Center Abstract Videoconferencing

More information

Network Model for Delay-Sensitive Traffic

Network Model for Delay-Sensitive Traffic Traffic Scheduling Network Model for Delay-Sensitive Traffic Source Switch Switch Destination Flow Shaper Policer (optional) Scheduler + optional shaper Policer (optional) Scheduler + optional shaper cfla.

More information

Multimedia Networking and Quality of Service

Multimedia Networking and Quality of Service Multimedia Networking and Quality of Service Mario Baldi Politecnico di Torino (Technical Univeristy of Torino) Department of Computer Engineering mario.baldi [at] polito.it +39 011 564 7067 staff.polito.it/mario.baldi

More information

Multimedia Applications over Packet Networks

Multimedia Applications over Packet Networks Multimedia Networking and Quality of Service Mario Baldi Technical Univeristy of Torino Computer Engineering Department mario.baldi@polito.it +39 011 564 7067 staff.polito.it/mario.baldi Nota di Copyright

More information

different problems from other networks ITU-T specified restricted initial set Limited number of overhead bits ATM forum Traffic Management

different problems from other networks ITU-T specified restricted initial set Limited number of overhead bits ATM forum Traffic Management Traffic and Congestion Management in ATM 3BA33 David Lewis 3BA33 D.Lewis 2007 1 Traffic Control Objectives Optimise usage of network resources Network is a shared resource Over-utilisation -> congestion

More information

Comparison of Shaping and Buffering for Video Transmission

Comparison of Shaping and Buffering for Video Transmission Comparison of Shaping and Buffering for Video Transmission György Dán and Viktória Fodor Royal Institute of Technology, Department of Microelectronics and Information Technology P.O.Box Electrum 229, SE-16440

More information

Scheduling Algorithms to Minimize Session Delays

Scheduling Algorithms to Minimize Session Delays Scheduling Algorithms to Minimize Session Delays Nandita Dukkipati and David Gutierrez A Motivation I INTRODUCTION TCP flows constitute the majority of the traffic volume in the Internet today Most of

More information

UBR Congestion controlled Video Transmission over ATM Eltayeb Omer Eltayeb, Saudi Telecom Company

UBR Congestion controlled Video Transmission over ATM Eltayeb Omer Eltayeb, Saudi Telecom Company UBR Congestion controlled Video Transmission over ATM Eltayeb Omer Eltayeb, Saudi Telecom Company ABSTRACT The ATM unspecified bit rate (UBR) class of service some times referred to as best effort service-

More information

Network-Adaptive Video Coding and Transmission

Network-Adaptive Video Coding and Transmission Header for SPIE use Network-Adaptive Video Coding and Transmission Kay Sripanidkulchai and Tsuhan Chen Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA 15213

More information

Module objectives. Integrated services. Support for real-time applications. Real-time flows and the current Internet protocols

Module objectives. Integrated services. Support for real-time applications. Real-time flows and the current Internet protocols Integrated services Reading: S. Keshav, An Engineering Approach to Computer Networking, chapters 6, 9 and 4 Module objectives Learn and understand about: Support for real-time applications: network-layer

More information

Quality of Service (QoS)

Quality of Service (QoS) Quality of Service (QoS) The Internet was originally designed for best-effort service without guarantee of predictable performance. Best-effort service is often sufficient for a traffic that is not sensitive

More information

RECOMMENDATION ITU-R BT.1720 *

RECOMMENDATION ITU-R BT.1720 * Rec. ITU-R BT.1720 1 RECOMMENDATION ITU-R BT.1720 * Quality of service ranking and measurement methods for digital video broadcasting services delivered over broadband Internet protocol networks (Question

More information

DigiPoints Volume 1. Student Workbook. Module 8 Digital Compression

DigiPoints Volume 1. Student Workbook. Module 8 Digital Compression Digital Compression Page 8.1 DigiPoints Volume 1 Module 8 Digital Compression Summary This module describes the techniques by which digital signals are compressed in order to make it possible to carry

More information

Prof. Dr. Abdulmotaleb El Saddik. site.uottawa.ca mcrlab.uottawa.ca. Quality of Media vs. Quality of Service

Prof. Dr. Abdulmotaleb El Saddik. site.uottawa.ca mcrlab.uottawa.ca. Quality of Media vs. Quality of Service Multimedia Communications Multimedia Technologies & Applications Prof. Dr. Abdulmotaleb El Saddik Multimedia Communications Research Laboratory School of Information Technology and Engineering University

More information

Fractional Lambda Switching

Fractional Lambda Switching Fractional Lambda Switching Mario Baldi and Yoram Ofek Synchrodyne Networks, Inc. 75 Maiden Lane, Suite 37 New York, NY0038 Abstract - Fractional Lambda (l) Switching (FlS ) adds the necessary efficiency

More information

Basics (cont.) Characteristics of data communication technologies OSI-Model

Basics (cont.) Characteristics of data communication technologies OSI-Model 48 Basics (cont.) Characteristics of data communication technologies OSI-Model Topologies Packet switching / Circuit switching Medium Access Control (MAC) mechanisms Coding Quality of Service (QoS) 49

More information

Asynchronous Transfer Mode (ATM) ATM concepts

Asynchronous Transfer Mode (ATM) ATM concepts Asynchronous Transfer Mode (ATM) Asynchronous Transfer Mode (ATM) is a switching technique for telecommunication networks. It uses asynchronous time-division multiplexing,[1][2] and it encodes data into

More information

Network Working Group Request for Comments: 1046 ISI February A Queuing Algorithm to Provide Type-of-Service for IP Links

Network Working Group Request for Comments: 1046 ISI February A Queuing Algorithm to Provide Type-of-Service for IP Links Network Working Group Request for Comments: 1046 W. Prue J. Postel ISI February 1988 A Queuing Algorithm to Provide Type-of-Service for IP Links Status of this Memo This memo is intended to explore how

More information

Module 6 STILL IMAGE COMPRESSION STANDARDS

Module 6 STILL IMAGE COMPRESSION STANDARDS Module 6 STILL IMAGE COMPRESSION STANDARDS Lesson 19 JPEG-2000 Error Resiliency Instructional Objectives At the end of this lesson, the students should be able to: 1. Name two different types of lossy

More information

CODING METHOD FOR EMBEDDING AUDIO IN VIDEO STREAM. Harri Sorokin, Jari Koivusaari, Moncef Gabbouj, and Jarmo Takala

CODING METHOD FOR EMBEDDING AUDIO IN VIDEO STREAM. Harri Sorokin, Jari Koivusaari, Moncef Gabbouj, and Jarmo Takala CODING METHOD FOR EMBEDDING AUDIO IN VIDEO STREAM Harri Sorokin, Jari Koivusaari, Moncef Gabbouj, and Jarmo Takala Tampere University of Technology Korkeakoulunkatu 1, 720 Tampere, Finland ABSTRACT In

More information

Interframe coding A video scene captured as a sequence of frames can be efficiently coded by estimating and compensating for motion between frames pri

Interframe coding A video scene captured as a sequence of frames can be efficiently coded by estimating and compensating for motion between frames pri MPEG MPEG video is broken up into a hierarchy of layer From the top level, the first layer is known as the video sequence layer, and is any self contained bitstream, for example a coded movie. The second

More information

Unit 2 Packet Switching Networks - II

Unit 2 Packet Switching Networks - II Unit 2 Packet Switching Networks - II Dijkstra Algorithm: Finding shortest path Algorithm for finding shortest paths N: set of nodes for which shortest path already found Initialization: (Start with source

More information

Lecture 17: Distributed Multimedia

Lecture 17: Distributed Multimedia 06-06798 Distributed Systems Lecture 17: Distributed Multimedia Distributed Systems 1 Overview Characteristics of multimedia systems audio, video, etc delivery in real time, on time Quality of service

More information

Congestion in Data Networks. Congestion in Data Networks

Congestion in Data Networks. Congestion in Data Networks Congestion in Data Networks CS420/520 Axel Krings 1 Congestion in Data Networks What is Congestion? Congestion occurs when the number of packets being transmitted through the network approaches the packet

More information

BROADBAND AND HIGH SPEED NETWORKS

BROADBAND AND HIGH SPEED NETWORKS BROADBAND AND HIGH SPEED NETWORKS ATM SERVICE CATEGORIES Service Categories : represent particular combinations of traffic parameters and QoS parameters. These parameters are determined by users for a

More information

Real-Time Protocol (RTP)

Real-Time Protocol (RTP) Real-Time Protocol (RTP) Provides standard packet format for real-time application Typically runs over UDP Specifies header fields below Payload Type: 7 bits, providing 128 possible different types of

More information

Maximizing the Number of Users in an Interactive Video-on-Demand System

Maximizing the Number of Users in an Interactive Video-on-Demand System IEEE TRANSACTIONS ON BROADCASTING, VOL. 48, NO. 4, DECEMBER 2002 281 Maximizing the Number of Users in an Interactive Video-on-Demand System Spiridon Bakiras, Member, IEEE and Victor O. K. Li, Fellow,

More information

Chapter 10. Circuits Switching and Packet Switching 10-1

Chapter 10. Circuits Switching and Packet Switching 10-1 Chapter 10 Circuits Switching and Packet Switching 10-1 Content Switched communication networks Circuit switching networks Circuit-switching concepts Packet-switching principles X.25 (mentioned but not

More information

Lecture 4 Wide Area Networks - Congestion in Data Networks

Lecture 4 Wide Area Networks - Congestion in Data Networks DATA AND COMPUTER COMMUNICATIONS Lecture 4 Wide Area Networks - Congestion in Data Networks Mei Yang Based on Lecture slides by William Stallings 1 WHAT IS CONGESTION? congestion occurs when the number

More information

Overview Computer Networking What is QoS? Queuing discipline and scheduling. Traffic Enforcement. Integrated services

Overview Computer Networking What is QoS? Queuing discipline and scheduling. Traffic Enforcement. Integrated services Overview 15-441 15-441 Computer Networking 15-641 Lecture 19 Queue Management and Quality of Service Peter Steenkiste Fall 2016 www.cs.cmu.edu/~prs/15-441-f16 What is QoS? Queuing discipline and scheduling

More information

What Is Congestion? Computer Networks. Ideal Network Utilization. Interaction of Queues

What Is Congestion? Computer Networks. Ideal Network Utilization. Interaction of Queues 168 430 Computer Networks Chapter 13 Congestion in Data Networks What Is Congestion? Congestion occurs when the number of packets being transmitted through the network approaches the packet handling capacity

More information

Outline Introduction MPEG-2 MPEG-4. Video Compression. Introduction to MPEG. Prof. Pratikgiri Goswami

Outline Introduction MPEG-2 MPEG-4. Video Compression. Introduction to MPEG. Prof. Pratikgiri Goswami to MPEG Prof. Pratikgiri Goswami Electronics & Communication Department, Shree Swami Atmanand Saraswati Institute of Technology, Surat. Outline of Topics 1 2 Coding 3 Video Object Representation Outline

More information

Lecture 17 Multimedia Transport Subsystem (Part 3)

Lecture 17 Multimedia Transport Subsystem (Part 3) CS 414 Multimedia Systems Design Lecture 17 Multimedia Transport Subsystem (Part 3) Klara Nahrstedt Spring 2010 Administrative MP2: deadline Monday, March 1, demos 5-7pm (sign up in class on Monday) HW1:

More information

Video Streaming Over Multi-hop Wireless Networks

Video Streaming Over Multi-hop Wireless Networks Video Streaming Over Multi-hop Wireless Networks Hao Wang Dept. of Computer Information System, Cameron University hwang@cameron.edu Andras Farago, Subbarayan Venkatesan Dept. of Computer Science, The

More information

Multimedia Standards

Multimedia Standards Multimedia Standards SS 2017 Lecture 5 Prof. Dr.-Ing. Karlheinz Brandenburg Karlheinz.Brandenburg@tu-ilmenau.de Contact: Dipl.-Inf. Thomas Köllmer thomas.koellmer@tu-ilmenau.de 1 Organisational issues

More information

DIGITAL TELEVISION 1. DIGITAL VIDEO FUNDAMENTALS

DIGITAL TELEVISION 1. DIGITAL VIDEO FUNDAMENTALS DIGITAL TELEVISION 1. DIGITAL VIDEO FUNDAMENTALS Television services in Europe currently broadcast video at a frame rate of 25 Hz. Each frame consists of two interlaced fields, giving a field rate of 50

More information

Worst-case Ethernet Network Latency for Shaped Sources

Worst-case Ethernet Network Latency for Shaped Sources Worst-case Ethernet Network Latency for Shaped Sources Max Azarov, SMSC 7th October 2005 Contents For 802.3 ResE study group 1 Worst-case latency theorem 1 1.1 Assumptions.............................

More information

MPEG4 VIDEO OVER PACKET SWITCHED CONNECTION OF THE WCDMA AIR INTERFACE

MPEG4 VIDEO OVER PACKET SWITCHED CONNECTION OF THE WCDMA AIR INTERFACE MPEG4 VIDEO OVER PACKET SWITCHED CONNECTION OF THE WCDMA AIR INTERFACE Jamil Y. Khan 1, Pratik Das 2 School of Electrical Engineering and Computer Science, University of Newcastle, Callaghan, NSW 238,

More information

Wireless Networks. Communication Networks

Wireless Networks. Communication Networks Wireless Networks Communication Networks Types of Communication Networks Traditional Traditional local area network (LAN) Traditional wide area network (WAN) Higher-speed High-speed local area network

More information

SIMULATION OF PACKET DATA NETWORKS USING OPNET

SIMULATION OF PACKET DATA NETWORKS USING OPNET SIMULATION OF PACKET DATA NETWORKS USING OPNET Nazy Alborz, Maryam Keyvani, Milan Nikolic, and Ljiljana Trajkovic * School of Engineering Science Simon Fraser University Vancouver, British Columbia, Canada

More information

Advantages and disadvantages

Advantages and disadvantages Advantages and disadvantages Advantages Disadvantages Asynchronous transmission Simple, doesn't require synchronization of both communication sides Cheap, timing is not as critical as for synchronous transmission,

More information

Performance Analysis of Storage-Based Routing for Circuit-Switched Networks [1]

Performance Analysis of Storage-Based Routing for Circuit-Switched Networks [1] Performance Analysis of Storage-Based Routing for Circuit-Switched Networks [1] Presenter: Yongcheng (Jeremy) Li PhD student, School of Electronic and Information Engineering, Soochow University, China

More information

Frame Relay. Frame Relay: characteristics

Frame Relay. Frame Relay: characteristics Frame Relay Andrea Bianco Telecommunication Network Group firstname.lastname@polito.it http://www.telematica.polito.it/ Network management and QoS provisioning - 1 Frame Relay: characteristics Packet switching

More information

Scheduling. Scheduling algorithms. Scheduling. Output buffered architecture. QoS scheduling algorithms. QoS-capable router

Scheduling. Scheduling algorithms. Scheduling. Output buffered architecture. QoS scheduling algorithms. QoS-capable router Scheduling algorithms Scheduling Andrea Bianco Telecommunication Network Group firstname.lastname@polito.it http://www.telematica.polito.it/ Scheduling: choose a packet to transmit over a link among all

More information

Implementation of a leaky bucket module for simulations in NS-3

Implementation of a leaky bucket module for simulations in NS-3 Implementation of a leaky bucket module for simulations in NS-3 P. Baltzis 2, C. Bouras 1,2, K. Stamos 1,2,3, G. Zaoudis 1,2 1 Computer Technology Institute and Press Diophantus Patra, Greece 2 Computer

More information

Chapter -6 IMPROVED CONGESTION CONTROL MECHANISM FOR REAL TIME DATA TRANSMISSION

Chapter -6 IMPROVED CONGESTION CONTROL MECHANISM FOR REAL TIME DATA TRANSMISSION Chapter -6 IMPROVED CONGESTION CONTROL MECHANISM FOR REAL TIME DATA TRANSMISSION Chapter 6 IMPROVED CONGESTION CONTROL MECHANISM FOR REAL TIME DATA TRANSMISSION 6.1 Introduction Supporting Quality of Service

More information

3. Quality of Service

3. Quality of Service 3. Quality of Service Usage Applications Learning & Teaching Design User Interfaces Services Content Process ing Security... Documents Synchronization Group Communi cations Systems Databases Programming

More information

Quality Differentiation with Source Shaping and Forward Error Correction

Quality Differentiation with Source Shaping and Forward Error Correction Quality Differentiation with Source Shaping and Forward Error Correction György Dán and Viktória Fodor KTH, Royal Institute of Technology, Department of Microelectronics and Information Technology, {gyuri,viktoria}@imit.kth.se

More information

CSCD 433/533 Advanced Networks Spring Lecture 22 Quality of Service

CSCD 433/533 Advanced Networks Spring Lecture 22 Quality of Service CSCD 433/533 Advanced Networks Spring 2016 Lecture 22 Quality of Service 1 Topics Quality of Service (QOS) Defined Properties Integrated Service Differentiated Service 2 Introduction Problem Overview Have

More information

ATM Quality of Service (QoS)

ATM Quality of Service (QoS) ATM Quality of Service (QoS) Traffic/Service Classes, Call Admission Control Usage Parameter Control, ABR Agenda Introduction Service Classes and Traffic Attributes Traffic Control Flow Control Special

More information

Audio and video compression

Audio and video compression Audio and video compression 4.1 introduction Unlike text and images, both audio and most video signals are continuously varying analog signals. Compression algorithms associated with digitized audio and

More information

Communication Networks

Communication Networks Communication Networks Chapter 3 Multiplexing Frequency Division Multiplexing (FDM) Useful bandwidth of medium exceeds required bandwidth of channel Each signal is modulated to a different carrier frequency

More information

Module 10 MULTIMEDIA SYNCHRONIZATION

Module 10 MULTIMEDIA SYNCHRONIZATION Module 10 MULTIMEDIA SYNCHRONIZATION Lesson 36 Packet architectures and audio-video interleaving Instructional objectives At the end of this lesson, the students should be able to: 1. Show the packet architecture

More information

CS 344/444 Computer Network Fundamentals Final Exam Solutions Spring 2007

CS 344/444 Computer Network Fundamentals Final Exam Solutions Spring 2007 CS 344/444 Computer Network Fundamentals Final Exam Solutions Spring 2007 Question 344 Points 444 Points Score 1 10 10 2 10 10 3 20 20 4 20 10 5 20 20 6 20 10 7-20 Total: 100 100 Instructions: 1. Question

More information

Reduction of Periodic Broadcast Resource Requirements with Proxy Caching

Reduction of Periodic Broadcast Resource Requirements with Proxy Caching Reduction of Periodic Broadcast Resource Requirements with Proxy Caching Ewa Kusmierek and David H.C. Du Digital Technology Center and Department of Computer Science and Engineering University of Minnesota

More information

IP Differentiated Services

IP Differentiated Services Course of Multimedia Internet (Sub-course Reti Internet Multimediali ), AA 2010-2011 Prof. 7. IP Diffserv introduction Pag. 1 IP Differentiated Services Providing differentiated services in IP networks

More information

Wireless Networks (CSC-7602) Lecture 8 (15 Oct. 2007)

Wireless Networks (CSC-7602) Lecture 8 (15 Oct. 2007) Wireless Networks (CSC-7602) Lecture 8 (15 Oct. 2007) Seung-Jong Park (Jay) http://www.csc.lsu.edu/~sjpark 1 Today Wireline Fair Schedulling Why? Ideal algorithm Practical algorithms Wireless Fair Scheduling

More information

White Paper Digital Video Networks & Quality of Service

White Paper Digital Video Networks & Quality of Service White Paper Digital Networks & Quality of Service TABLE OF CONTENTS INTRODUCTION... 3 WHAT IS QUALITY OF SERVICE?... 3 DIGITAL VIDEO TRANSPORT REQUIREMENTS... 3 NETWORK TECHNOLOGIES... 4 INTEGRATED SERVICES

More information

1188 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 13, NO. 5, OCTOBER Wei Sun, Student Member, IEEE, and Kang G. Shin, Fellow, IEEE

1188 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 13, NO. 5, OCTOBER Wei Sun, Student Member, IEEE, and Kang G. Shin, Fellow, IEEE 1188 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 13, NO. 5, OCTOBER 2005 End-to-End Delay Bounds for Traffic Aggregates Under Guaranteed-Rate Scheduling Algorithms Wei Sun, Student Member, IEEE, and Kang

More information

Chapter 1. Introduction

Chapter 1. Introduction Chapter 1 Introduction In a packet-switched network, packets are buffered when they cannot be processed or transmitted at the rate they arrive. There are three main reasons that a router, with generic

More information

Advanced Mechanisms for Available Rate Usage in ATM and Differentiated Services Networks

Advanced Mechanisms for Available Rate Usage in ATM and Differentiated Services Networks Advanced Mechanisms for Available Rate Usage in ATM and Differentiated Services Networks Roland Bless, Dirk Holzhausen, Hartmut Ritter, Klaus Wehrle Institute of Telematics, University of Karlsruhe Zirkel

More information

ECE 417 Guest Lecture Video Compression in MPEG-1/2/4. Min-Hsuan Tsai Apr 02, 2013

ECE 417 Guest Lecture Video Compression in MPEG-1/2/4. Min-Hsuan Tsai Apr 02, 2013 ECE 417 Guest Lecture Video Compression in MPEG-1/2/4 Min-Hsuan Tsai Apr 2, 213 What is MPEG and its standards MPEG stands for Moving Picture Expert Group Develop standards for video/audio compression

More information

Episode 5. Scheduling and Traffic Management

Episode 5. Scheduling and Traffic Management Episode 5. Scheduling and Traffic Management Part 2 Baochun Li Department of Electrical and Computer Engineering University of Toronto Keshav Chapter 9.1, 9.2, 9.3, 9.4, 9.5.1, 13.3.4 ECE 1771: Quality

More information

Efficient support for interactive operations in multi-resolution video servers

Efficient support for interactive operations in multi-resolution video servers Multimedia Systems 7: 241 253 (1999) Multimedia Systems c Springer-Verlag 1999 Efficient support for interactive operations in multi-resolution video servers Prashant J. Shenoy, Harrick M. Vin Distributed

More information

Ch. 4 - WAN, Wide Area Networks

Ch. 4 - WAN, Wide Area Networks 1 X.25 - access 2 X.25 - connection 3 X.25 - packet format 4 X.25 - pros and cons 5 Frame Relay 6 Frame Relay - access 7 Frame Relay - frame format 8 Frame Relay - addressing 9 Frame Relay - access rate

More information

Fundamentals of Video Compression. Video Compression

Fundamentals of Video Compression. Video Compression Fundamentals of Video Compression Introduction to Digital Video Basic Compression Techniques Still Image Compression Techniques - JPEG Video Compression Introduction to Digital Video Video is a stream

More information

Multimedia Networking. Network Support for Multimedia Applications

Multimedia Networking. Network Support for Multimedia Applications Multimedia Networking Network Support for Multimedia Applications Protocols for Real Time Interactive Applications Differentiated Services (DiffServ) Per Connection Quality of Services Guarantees (IntServ)

More information

ADVANCED COMPUTER NETWORKS

ADVANCED COMPUTER NETWORKS ADVANCED COMPUTER NETWORKS Congestion Control and Avoidance 1 Lecture-6 Instructor : Mazhar Hussain CONGESTION CONTROL When one part of the subnet (e.g. one or more routers in an area) becomes overloaded,

More information

ADAPTIVE PICTURE SLICING FOR DISTORTION-BASED CLASSIFICATION OF VIDEO PACKETS

ADAPTIVE PICTURE SLICING FOR DISTORTION-BASED CLASSIFICATION OF VIDEO PACKETS ADAPTIVE PICTURE SLICING FOR DISTORTION-BASED CLASSIFICATION OF VIDEO PACKETS E. Masala, D. Quaglia, J.C. De Martin Λ Dipartimento di Automatica e Informatica/ Λ IRITI-CNR Politecnico di Torino, Italy

More information

Testing Policing in ATM Networks

Testing Policing in ATM Networks Testing Policing in ATM Networks Policing is one of the key mechanisms used in ATM (Asynchrous Transfer Mode) networks to avoid network congestion. The HP E4223A policing and traffic characterization test

More information

UNIT 2 TRANSPORT LAYER

UNIT 2 TRANSPORT LAYER Network, Transport and Application UNIT 2 TRANSPORT LAYER Structure Page No. 2.0 Introduction 34 2.1 Objective 34 2.2 Addressing 35 2.3 Reliable delivery 35 2.4 Flow control 38 2.5 Connection Management

More information

How to achieve low latency audio/video streaming over IP network?

How to achieve low latency audio/video streaming over IP network? February 2018 How to achieve low latency audio/video streaming over IP network? Jean-Marie Cloquet, Video Division Director, Silex Inside Gregory Baudet, Marketing Manager, Silex Inside Standard audio

More information

15: OS Scheduling and Buffering

15: OS Scheduling and Buffering 15: OS Scheduling and ing Mark Handley Typical Audio Pipeline (sender) Sending Host Audio Device Application A->D Device Kernel App Compress Encode for net RTP ed pending DMA to host (~10ms according to

More information

Configuring QoS. Understanding QoS CHAPTER

Configuring QoS. Understanding QoS CHAPTER 29 CHAPTER This chapter describes how to configure quality of service (QoS) by using automatic QoS (auto-qos) commands or by using standard QoS commands on the Catalyst 3750 switch. With QoS, you can provide

More information

Digital Asset Management 5. Streaming multimedia

Digital Asset Management 5. Streaming multimedia Digital Asset Management 5. Streaming multimedia 2015-10-29 Keys of Streaming Media Algorithms (**) Standards (*****) Complete End-to-End systems (***) Research Frontiers(*) Streaming... Progressive streaming

More information

Achieving Distributed Buffering in Multi-path Routing using Fair Allocation

Achieving Distributed Buffering in Multi-path Routing using Fair Allocation Achieving Distributed Buffering in Multi-path Routing using Fair Allocation Ali Al-Dhaher, Tricha Anjali Department of Electrical and Computer Engineering Illinois Institute of Technology Chicago, Illinois

More information

What Is Congestion? Effects of Congestion. Interaction of Queues. Chapter 12 Congestion in Data Networks. Effect of Congestion Control

What Is Congestion? Effects of Congestion. Interaction of Queues. Chapter 12 Congestion in Data Networks. Effect of Congestion Control Chapter 12 Congestion in Data Networks Effect of Congestion Control Ideal Performance Practical Performance Congestion Control Mechanisms Backpressure Choke Packet Implicit Congestion Signaling Explicit

More information

BROADBAND AND HIGH SPEED NETWORKS

BROADBAND AND HIGH SPEED NETWORKS BROADBAND AND HIGH SEED NETWORKS LAYERS The function and associated information of the planes is as follows: The reference model is composed of the following planes: Control lane manages the call and connection.

More information

QoS-Aware IPTV Routing Algorithms

QoS-Aware IPTV Routing Algorithms QoS-Aware IPTV Routing Algorithms Patrick McDonagh, Philip Perry, Liam Murphy. School of Computer Science and Informatics, University College Dublin, Belfield, Dublin 4. {patrick.mcdonagh, philip.perry,

More information

Introduction to Real-Time Communications. Real-Time and Embedded Systems (M) Lecture 15

Introduction to Real-Time Communications. Real-Time and Embedded Systems (M) Lecture 15 Introduction to Real-Time Communications Real-Time and Embedded Systems (M) Lecture 15 Lecture Outline Modelling real-time communications Traffic and network models Properties of networks Throughput, delay

More information

A Hybrid Temporal-SNR Fine-Granular Scalability for Internet Video

A Hybrid Temporal-SNR Fine-Granular Scalability for Internet Video 318 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 11, NO. 3, MARCH 2001 A Hybrid Temporal-SNR Fine-Granular Scalability for Internet Video Mihaela van der Schaar, Member, IEEE, and

More information

CS244 Advanced Topics in Computer Networks Midterm Exam Monday, May 2, 2016 OPEN BOOK, OPEN NOTES, INTERNET OFF

CS244 Advanced Topics in Computer Networks Midterm Exam Monday, May 2, 2016 OPEN BOOK, OPEN NOTES, INTERNET OFF CS244 Advanced Topics in Computer Networks Midterm Exam Monday, May 2, 2016 OPEN BOOK, OPEN NOTES, INTERNET OFF Your Name: Answers SUNet ID: root @stanford.edu In accordance with both the letter and the

More information

Digital Video Processing

Digital Video Processing Video signal is basically any sequence of time varying images. In a digital video, the picture information is digitized both spatially and temporally and the resultant pixel intensities are quantized.

More information

Introduction to ATM Traffic Management on the Cisco 7200 Series Routers

Introduction to ATM Traffic Management on the Cisco 7200 Series Routers CHAPTER 1 Introduction to ATM Traffic Management on the Cisco 7200 Series Routers In the latest generation of IP networks, with the growing implementation of Voice over IP (VoIP) and multimedia applications,

More information

COMP 249 Advanced Distributed Systems Multimedia Networking. Multimedia Applications & User Requirements

COMP 249 Advanced Distributed Systems Multimedia Networking. Multimedia Applications & User Requirements COMP 249 Advanced Distributed Systems Multimedia Networking Multimedia Applications & User Requirements Kevin Jeffay Department of Computer Science University of North Carolina at Chapel Hill jeffay@cs.unc.edu

More information

A Preferred Service Architecture for Payload Data Flows. Ray Gilstrap, Thom Stone, Ken Freeman

A Preferred Service Architecture for Payload Data Flows. Ray Gilstrap, Thom Stone, Ken Freeman A Preferred Service Architecture for Payload Data Flows Ray Gilstrap, Thom Stone, Ken Freeman NASA Research and Engineering Network NASA Advanced Supercomputing Division NASA Ames Research Center Outline

More information

15 Data Compression 2014/9/21. Objectives After studying this chapter, the student should be able to: 15-1 LOSSLESS COMPRESSION

15 Data Compression 2014/9/21. Objectives After studying this chapter, the student should be able to: 15-1 LOSSLESS COMPRESSION 15 Data Compression Data compression implies sending or storing a smaller number of bits. Although many methods are used for this purpose, in general these methods can be divided into two broad categories:

More information

Chapter 1 Introduction

Chapter 1 Introduction Emerging multimedia, high-speed data, and imaging applications are generating a demand for public networks to be able to multiplex and switch simultaneously a wide spectrum of data rates. These networks

More information

Comparative Study of Partial Closed-loop Versus Open-loop Motion Estimation for Coding of HDTV

Comparative Study of Partial Closed-loop Versus Open-loop Motion Estimation for Coding of HDTV Comparative Study of Partial Closed-loop Versus Open-loop Motion Estimation for Coding of HDTV Jeffrey S. McVeigh 1 and Siu-Wai Wu 2 1 Carnegie Mellon University Department of Electrical and Computer Engineering

More information

MISB EG Motion Imagery Standards Board Engineering Guideline. 24 April Delivery of Low Bandwidth Motion Imagery. 1 Scope.

MISB EG Motion Imagery Standards Board Engineering Guideline. 24 April Delivery of Low Bandwidth Motion Imagery. 1 Scope. Motion Imagery Standards Board Engineering Guideline Delivery of Low Bandwidth Motion Imagery MISB EG 0803 24 April 2008 1 Scope This Motion Imagery Standards Board (MISB) Engineering Guideline (EG) provides

More information

Resource allocation in networks. Resource Allocation in Networks. Resource allocation

Resource allocation in networks. Resource Allocation in Networks. Resource allocation Resource allocation in networks Resource Allocation in Networks Very much like a resource allocation problem in operating systems How is it different? Resources and jobs are different Resources are buffers

More information

Video Compression An Introduction

Video Compression An Introduction Video Compression An Introduction The increasing demand to incorporate video data into telecommunications services, the corporate environment, the entertainment industry, and even at home has made digital

More information

Rate Control of Real-time MPEG-2 over ATM DBR Service with Bandwidth Re-negotiation

Rate Control of Real-time MPEG-2 over ATM DBR Service with Bandwidth Re-negotiation Rate Control of Real-time MPEG- over ATM DBR Service with Bandwidth Re-negotiation Kentarou Fukuda Naoki Wakamiya Masayuki Murata Hideo Miyahara Department of Informatics and Mathematical Science Graduate

More information

Resource Reservation Protocol

Resource Reservation Protocol 48 CHAPTER Chapter Goals Explain the difference between and routing protocols. Name the three traffic types supported by. Understand s different filter and style types. Explain the purpose of tunneling.

More information

Review and Implementation of DWT based Scalable Video Coding with Scalable Motion Coding.

Review and Implementation of DWT based Scalable Video Coding with Scalable Motion Coding. Project Title: Review and Implementation of DWT based Scalable Video Coding with Scalable Motion Coding. Midterm Report CS 584 Multimedia Communications Submitted by: Syed Jawwad Bukhari 2004-03-0028 About

More information

Real-Time ABR, MPEG2 Streams over VBR, and Virtual Source/Virtual Destination rt-abr switch

Real-Time ABR, MPEG2 Streams over VBR, and Virtual Source/Virtual Destination rt-abr switch Real-Time ABR, MPEG2 Streams over VBR, and Virtual Source/Virtual Destination rt-abr switch Professor of Computer and Information Sciences The Ohio State University Columbus OH 432101-1277 http://www.cis.ohio-state.edu/~jain/

More information

CHAPTER 3 EFFECTIVE ADMISSION CONTROL MECHANISM IN WIRELESS MESH NETWORKS

CHAPTER 3 EFFECTIVE ADMISSION CONTROL MECHANISM IN WIRELESS MESH NETWORKS 28 CHAPTER 3 EFFECTIVE ADMISSION CONTROL MECHANISM IN WIRELESS MESH NETWORKS Introduction Measurement-based scheme, that constantly monitors the network, will incorporate the current network state in the

More information

Configuring QoS. Finding Feature Information. Prerequisites for QoS. General QoS Guidelines

Configuring QoS. Finding Feature Information. Prerequisites for QoS. General QoS Guidelines Finding Feature Information, on page 1 Prerequisites for QoS, on page 1 Restrictions for QoS, on page 2 Information About QoS, on page 2 How to Configure QoS, on page 10 Monitoring Standard QoS, on page

More information

SINGLE PASS DEPENDENT BIT ALLOCATION FOR SPATIAL SCALABILITY CODING OF H.264/SVC

SINGLE PASS DEPENDENT BIT ALLOCATION FOR SPATIAL SCALABILITY CODING OF H.264/SVC SINGLE PASS DEPENDENT BIT ALLOCATION FOR SPATIAL SCALABILITY CODING OF H.264/SVC Randa Atta, Rehab F. Abdel-Kader, and Amera Abd-AlRahem Electrical Engineering Department, Faculty of Engineering, Port

More information

Flexible Bandwidth Provisioning in WDM Networks by Fractional Lambda Switching

Flexible Bandwidth Provisioning in WDM Networks by Fractional Lambda Switching Flexible Bandwidth Provisioning in WDM Networks by Fractional Lambda Switching Achille Pattavina, Donato Grieco Politecnico di Milano, Milano, Italy e-mail: pattavina@elet.polimi.it Yoram Ofek Synchrodyne

More information