# AUDIO AND VIDEO COMMUNICATION MEEC EXERCISES. (with abbreviated solutions) Fernando Pereira

Save this PDF as:

Size: px
Start display at page:

Download "AUDIO AND VIDEO COMMUNICATION MEEC EXERCISES. (with abbreviated solutions) Fernando Pereira"

## Transcription

1 AUDIO AND VIDEO COMMUNICATION MEEC EXERCISES (with abbreviated solutions) Fernando Pereira INSTITUTO SUPERIOR TÉCNICO Departamento de Engenharia Electrotécnica e de Computadores September 2014

2 1. Photographic Imaging 1.1) 2 nd Exam 1993/1994, 20 July 1994 Consider the transmission of digital images with a resolution of luminance samples and half this resolution, in each direction, for the chrominances (when used), using a 2 Mbit/s transmission channel. a) Considering that the transmission channel is available during 10 s, how many complete bilevel images can be transmited without any compression? b) And how many complete grey images can be transmited, in the same 10 s, if images with 128 grey levels are used (still without compression)? c) Considering now that a compression algorithm with compression factors of 20 and 15 for the luminance and chrominances, respectively, is used at 7 bit/sample, how many complete images can be transmited still in the same 10 s? 1.2) 2 nd Exam 1992/1993, 7 July 1993 a) As you know, most of the compression algorithms adopted by the JPEG standard are based on transform coding. Which are the main requirements for a transform to be used in the context of an image compression system? Why? b) Considering that the DCT is a linear transform, explain for which reason does Recommendation ITU-T H.261 compute the transform of the temporal differences (residuals) instead of the difference of the image transforms, considering that these two ways are mathematically equivalent. c) Which is the main reason justifying the Lohscheller matrices to be different for the luminance and chrominances? How does that fact impact the values in the Lohscheller matrices? 1

3 1.3) 1 st Exam 2011/2012, 11 June 2012 Consider the JPEG standard to code photographic images. a) Determine the average number of bits per pixel (considering both the luminance and the chrominances) that are spent when coding a 4:2:2 image with 16 bit/sample and a global compression factor (for the luminance and the chrominances) of 25. b) How many bits have to be spent to code a 4:2:0 colour image with luminance resolution if the luminance compression factor is 20 and the chrominance compression factor is twice the one for the luminance? c) Identify the simplest modulation that may be used to transmit in a 2 MHz bandwidth a 25 Hz video sequence coded as JPEG images in the format and conditions defined in b). 1.4) 2 nd Exam 1993/1994, 8 July 1994 Consider the JPEG compression standard for digital images. a) Determine the total transmission time for an ITU-R 601 format image ( luminance samples and samples for each chrominance with 8 bit/sample) coded with the sequential mode, considering that a 64 kbit/s channel is used and the compression factors are 15 and 20 for the luminance and chrominances, respectively. b) Considering now that the images are coded with the hierarchical mode, determine the transmission time for the 3 layers used assuming that: the transmission channel is the same. the spatial resolution for the base layer is luminance samples and samples for each chrominance. the spatial resolution doubles, in both directions, for each new layer. always 8 bit/samples. the compression factors for each layer increase 25 % for each new layer regarding the previous layer. the compression factors for the base layer are those indicated in a). c) State the relative benefits and drawbacks of using the two coding modes mentioned above, notably considering the transmission times computed. 2

4 1.5) 2 nd Exam 2005/2006, 8 July 2006 a) Determine the average number of bits per pixel used (considering the luminance and chrominances) when a 4:2:0 image with 8 bit/sample is coded with a global (luminance and chrominances) compression factor of 16. Determine the same metric if a compression factor of 20 is used for the luminance and a compression factor of 12 is used for the chrominances. b) What is the main difference between a lossless and a lossy image coding system? Which of these two types of systems is typically more important? Why? c) State a normative and a non-normative impact in terms of JPEG image compression from the fact that the human visual system is less sensitive to the higher frequencies than to the lower frequencies. d) Why is entropy coding used in most source coding systems, including JPEG codecs? What is the largest disadvantage of entropy coding for transmissions in mobile environments? 1.6) 1 st Exam 2005/2006, 20 June 2006 Assume that a user wants to access a database with JPEG coded images to search for some specific images. The maximum spatial resolution is for the luminance and for the chrominances, both with 8 bit/sample. a) Determine which JPEG coding modes have been used to code the images in the database if it is known that the users may access, in an efficient way, versions of the same image in several qualities and spatial resolutions. b) Assume that: 1. for sequential coding, the average compression factors for the luminance and chrominances are 10 and 15, respectively 2. for the base layer of the progressive coding mode, the compression factors are twice as high the factors for sequential coding 3. for the next layer, the compression factors are 3 times higher than for the sequential coding mode Knowing that, on average, each user quickly browses 4 images in the base layer before finding the target image, that the user takes 2 s to decide if an image is the target image or not, and that the transmission is made at 64 kbit/s, determine the total time required, on average, to get a target image in the maximum quality if images with 2 layers are used and all image layers are sequentially transmited unless the user stops the transmission by browsing to the next image. Finally, assume that the decoding times are negligible. c) Identify which would be the main consequences (at least 2) if the JPEG standard would have used a spatial transform with base functions not independent from the image to code. 3

5 1.7) 2 nd Exam 2011/2012, 29 June 2012 Consider the JPEG standard to code photographic images. a) Determine the compression factors that would be needed for the luminance and for the chrominances to spend an average number of 0.64 bit/pixel (considering both the luminance and the chrominances) when coding a 4:2:0 image with 8 bit/sample, knowing that the average luminance compression factor is twice the average chrominances compression factor. b) Determine the total number of bits that have to be spent to code a , 4:2:2, 8 bit/sample image if an average number of 3 DCT coefficients are coded per block and each coefficient costs, on average, 4 bits; additionally consider that the EOB (End of Block) word costs 2 bits and all blocks in the image spend bits. c) Consider a 4:2:2, 8 bit/sample image coded with the hierarchical mode. How many layers can we use to code the image if the base layer is for the luminance and has a global (luminance and chrominances) compression factor of 20, the global compression factor doubles for each new layer, each new layer has twice the resolution in both directions, and the total number of bits spent should be less than 10 6 bits. 1.8) 1 st Exam 2093/2004, 28 June 2004 Consider the JPEG standard to code digital images. a) Identify and explain the working process of the two JPEG ways of implementing the progressive coding mode. b) How would you select the prediction mode to use in the context of the JPEG lossless coding mode? What would you do if there was a need to guarantee that the effect of transmission errors do not propagate too much in the decoded image? c) Which are the main benefits and drawbacks of using the VLI codes to code the amplitude of the DCT coefficients in the JPEG baseline coding process? d) Explain the relevance of the concept of entropy of a source for the designer of a source encoder. 4

6 2. Videotelephony and Videoconference 2.1) 1 st Exam 1992/1993, 7 June 1993 Consider the specification of a video compression algorithm for digital videotelephony based on the 4 main tools used in Recommendation ITU-T H.261. Assume that the video data has a luminance spatial resolution at 10 Hz; chrominance spatial resolution is half the luminance spatial resolution in both directions. a) Considering that each luminance and chrominance sample is represented with 8 bits, compute the bitrate without compression needed for the transmission of this video data. b) To limit error propagation along the various decoded frames, 2 coding modes are created, depending on the coding tools used. From the 4 main available tools mentioned above, state and justify which tools can be used for each coding mode while achieving the identified target with the best compression performance. c) Compute the global compression factor assuming that for each image all the coding tools mentioned above are used with the following compression factors. Assume that the compression factors are independent in the sense that their effects may be accumulated. Coding Tool Luminance Compression Factor Chrominance Compression Factor ) 1 st Exam 1993/1994, 22 June 1994 As you know, Recommendation ITU-T H.261 is an international standard for the compression of video data in videotelephony and videoconference applications. a) State for which reason motion estimation and compensation are made at macroblock level (16 16 pixels). b) State the reason why the motion vectors are coded in a differential way. 5

7 c) Explain the motivation for the fact that no DCT coefficients selection thresholds are standardized for the quantization process. Which is the main advantage of this option? d) Explain the motivation for the fact that DC DCT coefficients are quantized differently from AC DCT coefficients. What is the difference? 2.3) 2 nd Exam 1993/1994, 8 July 1994 Consider the video compression algorithm for videotelephony and videoconference specified in Recommendation ITU-T H.261. For certain video sequences, the probabilities for the various macroblock (MB) coding classes were measured and the results in the table below were obtained. a) Indicate the set of codewords to code each MB coding class if Huffman entropy coding is used. b) Indicate 3 reasons that may justify the high percentage of macroblocks coded with the Intra mode. c) Which factors determine the choice of the maximum amplitude for the motion vectors components? Coding Type Coding Mode Probability A Intra 0,25 B Inter 0,6 C Inter + Motion Compensation 0,05 D Inter + Motion Compensation + Filter 0.05 E Inter + Filter 0,05 2.4) 2 nd Exam 1993/1994, 20 July 1994 Consider a videotelephony communication using Recommendation ITU-T H.261 at a bitrate of 64 kbit/s. The video data is coded using a CIF spatial resolution at 10 Hz. a) Knowing that 12800, 3200 and bits were spent on the coding of the first, second and third images, respectively, compute (justifying) which are the minimum acceptable size for the encoder output buffer and the initial visualization delay at the receiver, considering that the sequence of images mentioned above includes the worst case in terms of bit production. b) Indicate, justifying, which is the maximum size of the encoder output buffer if a maximum initial visualization delay of 200 ms is required (naturally the bits in a) are not relevant anymore here). 6

8 2.5) 1 st Exam 1994/1995, 30 June 1995 Consider Recommendation ITU-T H.261 for the coding of video data in videotelephony and videoconference applications. a) Indicate the 3 conditions for which the motion vector prediction used for motion vector differential coding is zero. Justify all of them. b) As you know, a rather efficient bitrate control solution for H.261 controls the quantization step as a function of the encoder output buffer fullness. Suggest a mathematical formula to express this dependency considering the values that the quantization step may take for this coded. Indicate 3 necessary conditions that this formula must fulfill. 2.6) 1st Exam 1995/1996, 26 June 1996 Consider a videotelephone communications using Recommendation ITU-T H.261 for video coding at a rate of 64 kbit/s. The video sequence is coded with a CIF spatial resolution and a frame rate of 12.5 Hz. Each video image to code is horizontally divided into two equal parts but while the bottom part is fixed, the top part is moving. Since the encoder processes sequentially the macroblocks, it is observed that all bits are uniformely generated in the first half of the time interval that the encoder usually dedicates to encode each image. At the encoder, the bits wait for transmission in an output buffer. Knowing that the first image has used bits, the second image bits, and the third image 2560 bits, determine: a) The time instant at which the receiver has obtained all bits for the first image. b) The minimum size of the encoder output buffer in order the bits mentioned above are appropriately transmitted. c) The minimum visualization delay to apply at the decoder assuming that the encoder output buffer is the one determined in b) and the encoder may generate the bits with any distribution in the interval between the acquisition of two images. 2.7) 2 nd Exam 2001/2002, 15 July 2002 a) Which is the global compression factor necessary to be able to transmit a video sequence with CIF spatial resolution ( and luminance and chrominance sample, respectively) at 10 Hz in a ISDN channel with 64 kbit/s, knowing that 10% of the overall available rate is used for synchronization and multiplexing data? b) Knowing that H.261 codes the quantization step with a fixed length code and supposing that the quantization step is no more send for all GOBs but only once per frame, which would be the bitrate saved for a 10 Hz CIF video sequence (do not consider the fact that the quantization step may also be sent at macroblock level)? c) Indicate the two main reasons justifying that the most used rate control solution for H.261 is the variation of the quantization step. 7

9 2.8) 1 st Exam 2005/2006, 20 June 2006 Consider the laboratory session about Recommendation ITU-T H.261. a) In the time instants with higher video activity, notably at scene cuts, some macroblocks were classified in a particularly indesirable way in terms of video decoded quality. What was this coding mode and why was it used? What type of coding does it imply? What distinguishes these macroblocks from others using the same H.261 coding mode without the same problems in terms of video quality and what is the subjective effect typically associated to these macroblocks? b) The existence of an encoder output buffer in H.261 codecs has, at least, two important impacts in terms of the quality of service provided to the final user; one impact is positive while the other is negative. What are these impacts and why do they happen? c) Assuming that an encoder is having difficulties to work in real-time, identify two possible ways to address these difficulties while impacting as less as possible the final quality offered to the users. 2.9) 1 st Exam 2011/2012, 11 June 2012 Consider a videotelephony communication using Recommendation ITU-T H.261. The video sequence is coded with a CIF spatial resolution and a frame rate of 12.5 Hz at a rate of 128 kbit/s. The video content to code is horizontally divided into two equal parts; however, while the bottom part is fixed, the top part is moving. Since the encoder processes sequentially the macroblocks, it is observed that all bits are uniformly generated in the first half of the time interval that the encoder usually dedicates to encode each image. At the encoder, the bits wait for transmission in an output buffer. Knowing that the first image has used bits, the second image bit, and the third image bits, determine: a) The time instants at which the receiver obtains all bits for the first, second and third images. b) The minimum size of the encoder output buffer in order all bits above are transmitted without problems. c) The initial visualization delay associated to the system defined in b). d) The maximum number of bits that the 4th image may spent (still assuming that it only spends bits in the top half). 8

10 2.10) 1 st Exam 2008/2009, 24 June 2009 Consider a ITU-T H.261 videophone system coding video with a spatial resolution of pixels for the luminance, 4:2:0 chrominance subsampling, with 8 bit/sample), at 12.5 Hz. Assume that the average compression factor (measured over all the macroblocks in the image), without any external constraints in terms of coding modes, is 25 for the luminance and 30 for the chrominances (header bits not included). Assume that 500 bits of overhead are spent per frame. a) Assuming that for each frame, on average, only 200 macroblocks generate code bits (the remaining ones are so similar to the previous image that no update is needed), determine the average overall compression factor (including luminance and chrominances) measured over the macroblocks which effectively generate code bits, including also the overhead bits. b) If for editing reasons, all the macroblocks of all frames had to be coded in intra mode, what would be the total bitrate assuming that the compression factor for this type of coding is half the compression factors indicated for the luminance and chrominance? c) For the situation in a), assume that to guarantee a higher error protection, one out of each 100 macroblocks spending bits is necessarily coded in intra mode. Assuming that the intra coding mode has a compression factor as defined in b), determine what would be the global compression factor (considering all the macroblocks in the image) corresponding to this situation also including the header bits. d) If a transmission rate of 1 Mbit/s is used, what would be maximum number of bits that the first frame may spend if a maximum acquisition-visualization delay of 200 ms would be requested. Assume that the encoder generates the bits for each frame uniformly in the time period between the acquisition moments of each two successive frames. 9

11 3. Digital Video Storage 3.1) 1 st Exam 1992/1993, 7 June 1993 Consider the ISO/IEC MPEG-1 Video coding standard. a) In terms of coding tools, explain the fundamental differences between this algorithm and the one specified in Recommendation ITU-T H.261. b) Explain which are the factors determining the selection of the N and M temporal prediction structure parameters for MPEG-1 video coding. c) Explain the reason why both the H.261 and MPEG-1 Video coding algorithms use relative macroblock addressing within the GOBs or slices (with the exception of the first macroblock transmitted). 3.2) 2 nd Exam 1993/1994, 20 July 1994 Consider the MPEG-1 standard for digital video storage. a) Indicate the reason why it is not advisable to uniformly distribute the available bits by the various types of frames defined in this standard. b) Indicate, justifying, which is the main characteristic of the video decoder which dimensioning is strongly influenced by the implementation of the normal reverse mode? Explain why the same does not typically happen with the fast reverse mode. c) Consider using the MPEG-1 Video coding algorithm at 25 Hz with M= 3 and N= 12 characterizing the temporal coding structure. If I frames get 3 times more bits than P frames and P frames get 4 times more bits than B frames (always on average), determine the bitrate for the video data if, on average, each B frame macroblock uses 50 bits (each frame has 396 macroblocks). 10

12 3.3) 2 nd Exam1992/1993, 7 July 1993 The MPEG-1 Video standard basically uses the same coding tools used by Recommendation ITU- T H.261 and has a target bitrate falling within the bitrate range of H.261 (about 1.1 Mbit/s for video). a) In this situation, which reasons motivated the specification of the MPEG-1 Video standard instead of simply using the H.261 solution already available? b) In terms of motion estimation and compensation, the MPEG-1 Video standard includes two big novelties regarding H.261. Which are these novelties? c) Consider using the MPEG-1 Video coding algorithm with 1.15 Mbit/s at 25 Hz with M= 4 and N= 16 characterizing the temporal coding structure. If I frames get 4 times more bits than P frames and P frames get 5 times more bits than B frames (always on average), compute the average number of bits available per macroblock for each type of frames assuming that each frame has 396 macroblocks. 3.4) 2 nd Exam 1998/1999, 28 July 1999 Suppose that you are contacted to project a high quality videoconference system for a multinational company. The maximum acceptable acquisition-visualization delay is 250 ms. The spatial resolution is (Y) and (Cr, Cb) at 25 Hz (8 bit/sample). Assume that you have available, providing the necessary video quality, two alternative systems: 1) a MPEG-1 Video codec with M=3 and N=3; 2) a H.261 codec. The average compression factors for the various frame types in the MPEG-1 Video codec are indicated in the table below. The average H.261 compression factors are the same as the MPEG-1 Video compression factors for the P frames. The first frame in H.261 coding has a compression factor similar to the MPEG-1 Video I frames and no P frame uses more bits than these frames. Frame Type Luminance CompressionFactor Chrominance CompressionFactor I P B Indicate, justifying, which solution would you propose to your client to satisfy his/her needs knowing that he/she intends to minimize the transmission costs. 3.5) 2 nd Exam 1997/1998, 6 July 1998 Suppose that you are contacted by a music clips production company which intends to start using digital systems for video processing and storage. The company tells you that it pretends a system with large flexibility in terms of edition maximum access time for each frame lower than 1 second and that it pretends to store the highest possible number of clips with 4 minutes in a disc 11

13 with 10 GBytes. The disc reading speed is 2 Mbit/s (constant). The clips have ITU-R 601 (European) resolution this means (Y) and (Cr, Cb) at 25 Hz. Assuming that you have available: 1. A JPEG codec providing compression factors of 10 and 15 (worst case) for the luminance and chrominances, respectively, with acceptable quality. 2. A MPEG-1 Video codec using I, P and B frames with a periodic, regular and symmetric temporal coding structure, providing compression factors of 12 and 15 for the luminance and chrominances, respectively, for the I frames, and compression factors of 25 and 30 for the luminance and chrominances, respectively, for the P frames (always worst cases). B frames have compression factors three times higher than I frames but to obtain those compression factors it is necessary to use coding anchors always immediately adjacent in time to the B frames. Indicate, justifying, which solution would propose to your client to satisfy his/her needs and state how many complete clips would you be able to store in the available disc. 3.6) 1 st Exam, 2001/2002, 26 June 2002 Consider the MPEG-1 Audio standard. a) Indicate three important parameters for the definition of the cost of a terminal for which the three layers of the MPEG-1 Audio standard offer a different trade-off. b) What is the utility of the psychoacoustic model for the audio coding solutions specified by the MPEG-1 Audio standard? c) Indicate the two main coding tools differences between the first/second layers and the third layer of the MPEG-1 Audio coding standard. 3.7) 2 nd Exam 2002/2003, 11 July 2003 Consider the MPEG-1 Audio standard. a) Explain what the thresholds of hearing and pain are. Which is their impact in audio coding? b) Explain what is the masking effect in audio coding. Explain the difference between temporal and frequency masking. c) What is the basic idea behind perceptual audio coding? Why are the characteristics of the auditory human system explored here and not the characteristics of the vocal tract like in speech coding? d) Indicate and explain the two main forms of audio coding in the frequency domain. 12

14 3.8) 2 nd Exam 2001/2002, 15 July 2002 a) Indicate five factors which are important for the selection of the type of storage support for digital video storage. b) Which are the two main temporal informations transmitted in a MPEG-1 Systems stream? Explain for which frame coding type these two temporal informations are different? c) Which was the target quality for the MPEG-1 video storage systems? Why? d) In which way and why are motion vectors coded in the MPEG-1 Video standard? e) Why does the MPEG-1 Video standard indicate at the frame level the variation range of the motion vectors? f) Explain the reason which motivated the MPEG-1 Video standard to adopt different order for the acquisition, transmission and visualization of the video data. 3.9) 2 nd Exam 2004/2005, 22 July 2005 Consider the MPEG-1 Audio standard. a) Determine how many complete stereo songs with a 22 khz bandwidth and 16 bit/sample with duration of 3 minutes is possible to store in a disk with 200 Mbytes if the coding is performed with MPEG-1 Audio Layer 3 to reach CD transparent quality. b) Why does this standard use the DCT with overlapping window? c) What is the objective to use a scale factor for each audio subband? What would happen if the coding process would not use these scale factors? What is the main difference between Layers 1 and 2 in terms of the coding of these scale factors? d) Which main factors would you take into account (at least 3) to select one of the MPEG-1 Audio Layers to code the audio for a certain application? 3.10) 1 st Exam 2011/2012, 29 June 2012 Consider the MPEG-1 Audio standard to code audio content with 22 khz bandwidth; assume reasonable compression factors and the most usual number of bits per sample. a) How many complete stereo music pieces, with a duration of 4 minutes, can we store in a 900 MBytes disk using the Layer 3 of the MPEG-1 Audio standard to code the music content with a transparent quality regarding CD music content. b) What is the maximum duration of each music piece that we can afford if we want to store 1000 musics in the same disk as above using a Layer 2 MPEG-1 Audio codec? c) Explain how would the maximum number of stored musics vary if we increase the audio bandwidth three times but the audio becomes mono and not anymore stereo. d) Describe two main technical differences between the MPEG-1 Audio Layer 2 and Layer 3 codecs and the corresponding advantages. 13

15 3.11) 2 nd Exam 2009/2010, 12 July 2010 Consider using the MPEG-1 Video codec to code CIF (396 macroblocks) video information at 25 Hz to be stored in a CD. Assume that M=3 is used and the I get 3 times more bits than the P frames while the P frames get 4 times more bits than the B frames (always on average). The average number of bits per macroblock in a B frame is 50. For the conditions above, determine the acceptable set of N values if it is requested that the video bitrate does not exceed 1.8 Mbit/s and the maximum access time does not exceed 400 ms. Assume that the reading rate is the same as the coding rate. 3.12) 2 nd Exam 2011/2012, 29 June 2012 Suppose that you have been contacted by a company to design a videoconference solution to work between the various EURO 2012 stadiums using the lowest possible bitrate while guaranteeing the necessary minimum video quality. The company also requires that the initial visualization delay (measured as the maximum time difference between corresponding acquisition and visualization instants) is below 300 ms. The video resolution is CIF ( luminance samples), 4:2:0 at 12.5 Hz with 8 bit/sample. Assume that you have available and providing the necessary minimum video quality two systems: 1. H.261 system with average compression factors of 18 and 22 for the luminance and chrominances, respectively, and critical compression factors (for the most difficult images) of 12 and 15 for the luminance and chrominances, respectively. 2. MPEG-1 Video system with N = M = 3 and average compression factors of a. 18 and 22 for the luminance and chrominances, respectively, in the I frames b. 25 and 35 for the luminance and chrominances, respectively, in the B and P frames The critical compression factors are 75 % of the average compression factors for all frame types. a) Determine which of the solutions above you would select to better satisfy the needs of your client assuming that the transmission rate for each solution correspond to its coding rate. b) Determine for which transmission bitrates the MPEG-1 solution would satisfy the initial visualization delay requirement, assuming that for this solution the transmission rate may be regulated (meaning that the coding and transmission rate would be different). c) Explain 2 performance impacts of increasing the value of M, this means increasing the number of B frames between two anchor frames. 14

16 4. Digital Television 4.1) 1 st Exam 1994/1995, 19 June 1995 Consider the implementation of a digital TV system with the ITU-R 601 spatial and temporal resolutions ( and samples for luminance and chrominances, respectively, at 25 Hz). a) Assuming that the transmission channel has a capacity of 100 Mbit/s and no compression algorithm is used, what is the maximum number of bits per sample that can be used to sample the luminance signal considering the luminance and chrominance sample use the same number of bits per sample. b) Assuming now that a compression algorithm is used, providing compression factors of 20 and 25 for the luminance and chrominances, respectively, indicate what is the capacity of the transmission channel mentioned above which will remain free if the number of bits per sample is ) 2 nd Exam 1997/1998, 24 July 1998 Suppose that your company has been contacted to design the video system feeding the giant screen at Sony Plaza at EXPO 98. The digital transmission will be in high definition (Y) and (Cr, Cb) at 25 Hz (8 bit/sample). Assume that you have available providing the necessary quality for each frame, a MPEG-2 Video codec reaching the compression factors indicated in the table below. To guarantee adequate random access, at least one frame has to be coded in Intra mode every 300 ms. Finally, to reach the compression factors in the table, no more than three B frames should be introduced consecutively and a P frame must always be present between two I frames. Assuming that the intention is to minimize the bitrate to reduce the transmission costs, determine: 15

17 Frame Type Luminance Compression Factor Chrominance Compression Factor I P a) The best M and N values characterizing the (regular) temporal coding structure of I, P and B frames to be adopted. b) The average bitrate associated to the coding structured determined above. B c) The initial visualization delay at the receiver assuming that the transmission is made at the rate determined above, the N value is the same and M=N, and the critical compression factors for the I frames (for the more difficult frames) are 10% lower than the average compression factors indicated in the table (they are the same for the other frame types). Assume also that the coding and decoding times are negligible. 4.3) 2 nd Exam 1998/1999, 28 July 1999 a) Explain the main difference between the MPEG-1 and MPEG-2 standards in terms of the content they are able to code. What justifies this difference? b) Explain which are the main reasons (2) motivating the specification of profiles and levels in the MPEG-2 Video standard. c) Explain how it would be classified, in terms of profiles and levels, a decoder which capacity is below (even if very close) a certain conformance point. Why? 4.4) 2 nd Exam 1997/1998, 6 July 1998 Consider the MPEG-2 standard for digital video coding. a) Which were the main reasons (2) leading to the specification of profiles and levels in the context of this standard? b) How and why is a bitstream classified in terms of profiles and levels if its characteristics are just slightly above a certain conformance point? c) What would happen if a TV radio transmitter emits MPEG-2 coded video compliant to and the receiver is compliant to How does the situation change if the application context is now a multimedia database with online and offline MPEG-2 video coding? 16

18 4.5) 1 st Exam 2001/2002, 26 June 2002 a) What is the rationale delimiting the frontier between the technologies standardized by MPEG and DVB through ETSI? Why? b) Indicate three important differences between the program and transport streams specified in the MPEG-2 Systems standard. c) How many different Program Association Tables typically exist in a MPEG-2 transport stream? Why? d) The MPEG-2 Video standard classifies each image to code as image-frame or image-field. What is the difference between these two image types? What was the motivation to define these two image types? In which conditions are these two image types typically used for interlaced content? e) What MPEG-2 Video conformance points did DVB adopted and with what objectives? What would happen if DVB had not adopted levels but just profiles in the context of MPEG-2 Video? 4.6) 2 nd Exam 2001/2002, 15 July 2002 a) Explain what measures the so-called coding rate in systems like DVB-T. b) Indicate what is the main idea used in the OFDM modulation adopted in DVB-T to reduce the number of mutually interfering modulated symbols. c) What is the purpose of the guard interval used in the DVB-T modulation system? What is conceptually the minimum value of this interval? d) Indicate, justifying, which of the COFDM DVB-T variants is more indicated to cover small areas. 4.7) 2 nd Exam 2003/2004, 19 July 2004 Consider the DVB standards. a) Explain how the hierarchical 64-QAM modulation is used with benefit regarding a nonhierarchical version and for which conditions there are particular advantages in the adoption of this type of modulation. b) Identify two relevant ways of combining the hierarchical 64-QAM modulation with the scalable coding methods available in MPEG-2 Video. c) Identify two relevant factors for the selection of the encryption method to be used in the conditional access module of a digital pay TV system. 17

19 4.8) 2 nd Exam 2005/2006, 8 July 2006 Consider a DVB-T digital TV system. a) Determine how many complete stereo songs with a 22 khz bandwidth and 16 bit/sample with duration of 3 minutes is possible to store in a disk with 200 Mbytes if the coding is performed with MPEG-1 Audio Layer 3 to reach CD transparent quality. b) How many bytes do you have to spend for the coding of a 4:2:0 colour video sequence with 30 minutes and a luminance spatial resolution at 8 bit/sample at 25 Hz if the luminance compression factor is 30 and the chrominances compression factors is twice that value? c) Knowing that a DVB-T solution may insert 5 Mbit/s of source bits in a 8 MHz bandwidth channel, determine what would be the bitrate that may be inserted if all the system parameters stay the same with the exception of the channel coding ratio that goes from ½ to 2/3 and the modulation that goes from QPSK to 64-QAM. d) Which solution would you select from 1) 64-QAM modulation and ½ channel coding ratio and 2) QPSK and 5/6 channel codimg ratio, with all other parameters the same, if the selection criterion was the maximization of source bitrate. Knowing that the solution was to be used in a satellite transmission, would you remain more or less confident on the solution? Why? 4.9) 2 nd Exam 2011/2012, 29 June 2012 Consider a DVB digital TV system. a) Knowing that a DVB solution may insert 10 Mbit/s of total bitrate in a 8 MHz bandwidth channel, determine what would be the source bitrate that may be inserted if all the system parameters stay the same with the exception of the channel coding ratio that goes from ½ to 1/3 and the modulation that goes from 8-PSK to 64-QAM. b) Why is it essential in a Single Frequency Network that the transmitters send the same data and do that well synchronized to transmit the same symbol at precisely the same time? How do the transmitters obtain the necessary time reference? c) What are the two main components of the channel coding solution in DVB-x2? d) What parameter can be used to tune the correction capability of the channel coding solution and what does this parameter express? e) What is the main reason justifying the availability of two channel coding block lengths in DVBx2? 18

20 4.10) 1 st Exam 2010/2011, 14 June 2011 Consider a digital television system as specified by the DVB-T standard. a) Determine the typical bitrate to code in MP3 a mono audio signal with a bandwidth of 24 khz when using the most common number of bits per sample. b) Determine the rate spent in channel error protection in a DVB system using a 4/6 coding rate for a 20 Mbit/s source rate when using a typical 8 MHz channel. c) Determine the ratio between the rates corresponding to the lower and higher layers that you may transmit when using hierarchical 64-QAM. d) Determine a reasonable rate spent to transmit the MPEG-2 Systems Program Association Table (PAT) knowing that each table spends 300 bits. e) What solution would you choose between a first solution using a coding rate of ½ and modulation 64-QAM and a second solution using a coding rate of 2/3 and modulation 16-PSK based on the source rate (consider that all other parameters stay the same)? Why? 19

21 5. Advanced Video Coding 5.1) 2 nd Exam 2005/2006, 8 July 2006 Consider the H.264/AVC video coding standard. a) Regarding the slide groups defined in H.264/AVC, describe two ways of taking benefit of this concept. b) Explain why this standard uses an additional 4 4 Hadamard transform for the luminance DC coefficients from the macroblocks coded with the Intra coding mode. c) Identify the main difference, in terms of coding standards, between deblocking filters inside and outside the prediction loop. 5.2) 1 st Exam 2005/2006, 20 June 2006 a) What is the main conceptual novelty adopted in the H.264/AVC coding standard for the coding of the Intra macroblocks? b) Explain what is the H.264/AVC Constrained Intra coding mode and why was defined. c) Identify three reasons in terms of temporal prediction that may justify the substantial increase in the H.264/AVC encoding complexity. 5.3) 2 nd Exam 2004/2005, 22 July 2005 a) DVB as decided to adopt the H.264/AVC video coding standard. Which have been the H.264/AVC profiles adopted and for which conditions. For what reason did DVB adopt two different H.264/AVC profiles? b) Identify three main factors which must have been considered by DVB when adopting the new H.264/AVC profiles in addition to the previously adopted MPEG-2 Video profile. c) Which are the main benefits (1) and drawbacks (1) of the H.264/AVC standard regarding the MPEG-2 Video standard? 20

22 5.4) 1 st Exam 2008/2009, 24 June 2009 Consider the scalable coding of video content. a) Explain in your own words the type of functionality that it is possible to obtain with a scalable coded stream (and it is not possible to obtain with a non-scalable coded stream). b) Indicate and justify what type of networks (at least 2) may take benefit from the usage of scalable coded video. c) Explain what type of scalability does not normally imply any onus in terms of compression efficiency. d) Indicate and justify the main advantage and drawback of using a scalable coding stream in comparison with an alternative simulcasting solution. 5.5) 2 nd Exam 2008/2009, 20 July 2009 Consider that you become one of the YouTube managers in terms of multimedia content. a) Indicate what disk capacity would you need to store clips with an average duration of 4 minutes if the video has a spatial resolution of luminance samples, 4:2:0 subsampling format, 25 Hz, the audio (stereo) a sampling frequency of 48 khz (using the typical number of bit/sample), knowing that the audio compression factor is the typical MP3 compression factor for MP3 (for transparent quality) and the video compression factor is twice the audio compression factors for all video components. b) How much would that capacity change (in percentage), if the spatial resolution would be reduced to half in both directions and the audio sampling frequency would become 40 khz also knowing that in that case all compression factors would be reduced by 25 %. c) For the situations above, explain what video coding standard would you use, indicating 2 major system requirements justifying your choice. d) If you were given the possibility to improve the overall user experience in a very significative way using 1% more of capacity, how would you invest this capacity and why? 21

23 6. 3D Video 6.1) 1 st Exam 2011/2012, 11 June 2012 As you know, 3D video is nowadays very popular. a) Identify and explain the two main ways of providing the user a 3D video experience. b) Define both and explain the difference between stereo and movement parallaxes. Which of these types of parallax may be present in a multiview video system with many views? Why? c) Explain what does a requirement on view-switching random access typically ask for. When is this type of requirement important for an user? d) Compute the typical bitrate for a video stereo pair when using the Multiview Video Coding (MVC) with standard resolution if the two views are coded with similar PSNR. Compute the same bitrate for a system with 10 views. 6.2) 2 nd Exam 2011/2012, 29 June 2012 Consider a 3D video system. a) Explain what is a frame compatible stereo format. Also explain the difference between a spatial multiplexing and a time multiplexing frame compatible stereo format. b) Explain why frame-compatible stereo video tends to have higher spatial frequency content characteristics. c) What is the most important new feature/tool of the Multiview Video Coding (MVC) standard regarding the H.264/AVC standard? How does it work? d) What is the implication of the backward compatibility requirement for the MVC standard? e) If the backward compatible view in a MVC stereo pair spends 2 Mbit/s, what is the minimum rate that the second view has to spend if a perceptual quality similar to stereo simulcasting with 2+2 Mbit/s had to be achieved? Why? 22

24 Solutions for the Exercises 1. Photographic Imaging 1.1 a) 48; b) 6; c) a) 1.28 bit/pixel; b) for 16 bit/sample; c) 64-PSK, 64-QAM 1.4 a) s; b) s, 4.84 s and s. 1.5 a) 0.75 bit/pixel and bit/pixel. 1.6 a) Progressive and hierarchical; b) s. 1.7 a) 25; 12.5; b) ; c) a) Spectral selection, successive approximations 2. Videotelephony and Videoconference 2.1 a) bit/s; c) a) bit and 550 ms; b) 6400 bit. 2.6 a) 240 ms; b) bit; c) 520 ms. 2.7 a) 211.2; b) 550 bit/s; c) Speed and granularity of the reaction. 2.8 c) Using fast motion estimation methods and reducing the size of the motion window search. 2.9 a) 120, 280 and 360 ms; b) bits; c) 240 ms; d) bits a) 13.23; b) Mbit/s; c) 25.94; d) bits 3. Digital Video Storage 3.1 a) Bidirectional motion compensation and half-pixel motion vectors accuracy. 3.2 a) Because the various types of frame have different compression capabilities and thus obtaining rather constant quality implies a non-uniform bitrate allocation; b) Memory; c) 1.32 Mbit/s. 3.3 a) To addresses new requirements; b) Bidirectional motion compensation and half-pel motion vectors accuracy; c) NB = 39.54; NP = ; NI = MPEG-1 Video. 23

25 3.5 MPEG-1 Video. 3.6 a) Quality, complexity and bitrate; c) MDCT transform and Huffman entropy coding. 3.8 a) Capacity, reading speed, access time, type of access, durability, mobility, cost; b) PTS and DTS. 3.9 a) 75; b) Reduction of the block effect; d) Required quality; acceptable complexity; acceptable delay; available bitrate a) 255; b) s; c) reduction to 2/ N= a) H.261; b) > Mbit/s 4. Digital Television 4.1 a) 4 bit/sample; b) 94.4 Mbit/s. 4.2 a) N=6, M=3; b) Mbit/s; c) 305 ms. 4.3 a) Progressive versus interlaced content; b) To provide interoperability with limited complexity; c) Profile and level immediately below. 4.4 a) Providing interoperability and limiting complexity; b) Profile and level immediately above; c) Decoder is not able to decode since it does not know some of the coding tools eventually used. 4.5 c) 1; e) e 4.6 a) Fraction of the total rate used for source data; d) 2k. 4.8 a) 75; b) Mbytes; c) 20 Mbit/s; d) Solution a) Mbit/s 4.10 a) 64 kbit/s; b) 10 Mbit/s; c) ½; d) 3-15 kbit/s; d) Solution 1 5. Advanced Video Coding 5.3 a) Main for SDTV and High for HDTV; b) Increase in the compression capacity; increase in complexity; compatibility with existing terminals; c) Better compression efficiency and higher complexity. 5.4 b) Mobile networks and the Internet; c) Temporal scalability. 5.5 a) bits; b) %; c) H.264/AVC; d) Metadata. 6. 3D Video 6.1 d) 3.4 Mbit/s and 15.5 Mbit/s 6.2 e) 500 kbit/s 24

### 10.2 Video Compression with Motion Compensation 10.4 H H.263

Chapter 10 Basic Video Compression Techniques 10.11 Introduction to Video Compression 10.2 Video Compression with Motion Compensation 10.3 Search for Motion Vectors 10.4 H.261 10.5 H.263 10.6 Further Exploration

### Digital Video Compression and H.261 Recommendation

Digital Video Compression and H.261 Recommendation Fernando Pereira Klagenfurt, Austria, October 2008 Video versus Images Still Image Services No strong temporal requirements; no realtime notion. Video

### MPEG-2. ISO/IEC (or ITU-T H.262)

MPEG-2 1 MPEG-2 ISO/IEC 13818-2 (or ITU-T H.262) High quality encoding of interlaced video at 4-15 Mbps for digital video broadcast TV and digital storage media Applications Broadcast TV, Satellite TV,

### Audio and video compression

Audio and video compression 4.1 introduction Unlike text and images, both audio and most video signals are continuously varying analog signals. Compression algorithms associated with digitized audio and

### Perceptual Coding. Lossless vs. lossy compression Perceptual models Selecting info to eliminate Quantization and entropy encoding

Perceptual Coding Lossless vs. lossy compression Perceptual models Selecting info to eliminate Quantization and entropy encoding Part II wrap up 6.082 Fall 2006 Perceptual Coding, Slide 1 Lossless vs.

### COMP 249 Advanced Distributed Systems Multimedia Networking. The Video Data Type Coding & Compression Basics

COMP 9 Advanced Distributed Systems Multimedia Networking The Video Data Type Coding & Compression Basics Kevin Jeffay Department of Computer Science University of North Carolina at Chapel Hill jeffay@cs.unc.edu

### Ch. 4: Video Compression Multimedia Systems

Ch. 4: Video Compression Multimedia Systems Prof. Ben Lee (modified by Prof. Nguyen) Oregon State University School of Electrical Engineering and Computer Science 1 Outline Introduction MPEG Overview MPEG

### 5LSE0 - Mod 10 Part 1. MPEG Motion Compensation and Video Coding. MPEG Video / Temporal Prediction (1)

1 Multimedia Video Coding & Architectures (5LSE), Module 1 MPEG-1/ Standards: Motioncompensated video coding 5LSE - Mod 1 Part 1 MPEG Motion Compensation and Video Coding Peter H.N. de With (p.h.n.de.with@tue.nl

### Lecture 4: Video Compression Standards (Part1) Tutorial 2 : Image/video Coding Techniques. Basic Transform coding Tutorial 2

Lecture 4: Video Compression Standards (Part1) Tutorial 2 : Image/video Coding Techniques Dr. Jian Zhang Conjoint Associate Professor NICTA & CSE UNSW COMP9519 Multimedia Systems S2 2006 jzhang@cse.unsw.edu.au

### What is multimedia? Multimedia. Continuous media. Most common media types. Continuous media processing. Interactivity. What is multimedia?

Multimedia What is multimedia? Media types +Text + Graphics + Audio +Image +Video Interchange formats What is multimedia? Multimedia = many media User interaction = interactivity Script = time 1 2 Most

### 15 Data Compression 2014/9/21. Objectives After studying this chapter, the student should be able to: 15-1 LOSSLESS COMPRESSION

15 Data Compression Data compression implies sending or storing a smaller number of bits. Although many methods are used for this purpose, in general these methods can be divided into two broad categories:

### Compressed-Domain Video Processing and Transcoding

Compressed-Domain Video Processing and Transcoding Susie Wee, John Apostolopoulos Mobile & Media Systems Lab HP Labs Stanford EE392J Lecture 2006 Hewlett-Packard Development Company, L.P. The information

### 2.4 Audio Compression

2.4 Audio Compression 2.4.1 Pulse Code Modulation Audio signals are analog waves. The acoustic perception is determined by the frequency (pitch) and the amplitude (loudness). For storage, processing and

### Multimedia. What is multimedia? Media types. Interchange formats. + Text +Graphics +Audio +Image +Video. Petri Vuorimaa 1

Multimedia What is multimedia? Media types + Text +Graphics +Audio +Image +Video Interchange formats Petri Vuorimaa 1 What is multimedia? Multimedia = many media User interaction = interactivity Script

### Mesh Based Interpolative Coding (MBIC)

Mesh Based Interpolative Coding (MBIC) Eckhart Baum, Joachim Speidel Institut für Nachrichtenübertragung, University of Stuttgart An alternative method to H.6 encoding of moving images at bit rates below

### Multimedia networked applications: standards, protocols and research trends

Multimedia networked applications: standards, protocols and research trends Maria Teresa Andrade FEUP / INESC Porto mandrade@fe.up.pt ; maria.andrade@inescporto.pt http://www.fe.up.pt/~mandrade/ ; http://www.inescporto.pt

### Mpeg 1 layer 3 (mp3) general overview

Mpeg 1 layer 3 (mp3) general overview 1 Digital Audio! CD Audio:! 16 bit encoding! 2 Channels (Stereo)! 44.1 khz sampling rate 2 * 44.1 khz * 16 bits = 1.41 Mb/s + Overhead (synchronization, error correction,

### Compression Part 2 Lossy Image Compression (JPEG) Norm Zeck

Compression Part 2 Lossy Image Compression (JPEG) General Compression Design Elements 2 Application Application Model Encoder Model Decoder Compression Decompression Models observe that the sensors (image

### Chapter 14 MPEG Audio Compression

Chapter 14 MPEG Audio Compression 14.1 Psychoacoustics 14.2 MPEG Audio 14.3 Other Commercial Audio Codecs 14.4 The Future: MPEG-7 and MPEG-21 14.5 Further Exploration 1 Li & Drew c Prentice Hall 2003 14.1

### DAB. Digital Audio Broadcasting

DAB Digital Audio Broadcasting DAB history DAB has been under development since 1981 at the Institut für Rundfunktechnik (IRT). In 1985 the first DAB demonstrations were held at the WARC-ORB in Geneva

### Multimedia Signals and Systems Motion Picture Compression - MPEG

Multimedia Signals and Systems Motion Picture Compression - MPEG Kunio Takaya Electrical and Computer Engineering University of Saskatchewan March 9, 2008 MPEG video coding A simple introduction Dr. S.R.

### FPGA based High Performance CAVLC Implementation for H.264 Video Coding

FPGA based High Performance CAVLC Implementation for H.264 Video Coding Arun Kumar Pradhan Trident Academy of Technology Bhubaneswar,India Lalit Kumar Kanoje Trident Academy of Technology Bhubaneswar,India

### Audio Coding and MP3

Audio Coding and MP3 contributions by: Torbjørn Ekman What is Sound? Sound waves: 20Hz - 20kHz Speed: 331.3 m/s (air) Wavelength: 165 cm - 1.65 cm 1 Analogue audio frequencies: 20Hz - 20kHz mono: x(t)

### CSCD 443/533 Advanced Networks Fall 2017

CSCD 443/533 Advanced Networks Fall 2017 Lecture 18 Compression of Video and Audio 1 Topics Compression technology Motivation Human attributes make it possible Audio Compression Video Compression Performance

### MPEG-4 General Audio Coding

MPEG-4 General Audio Coding Jürgen Herre Fraunhofer Institute for Integrated Circuits (IIS) Dr. Jürgen Herre, hrr@iis.fhg.de 1 General Audio Coding Solid state players, Internet audio, terrestrial and

### Chapter 2 MPEG Video Compression Basics

Chapter 2 MPEG Video Compression Basics B.G. Haskell and A. Puri 2.1 Video Coding Basics Video signals differ from image signals in several important characteristics. Of course the most important difference

### Advanced Encoding Features of the Sencore TXS Transcoder

Advanced Encoding Features of the Sencore TXS Transcoder White Paper November 2011 Page 1 (11) www.sencore.com 1.605.978.4600 Revision 1.0 Document Revision History Date Version Description Author 11/7/2011

### Chapter 11.3 MPEG-2. MPEG-2: For higher quality video at a bit-rate of more than 4 Mbps Defined seven profiles aimed at different applications:

Chapter 11.3 MPEG-2 MPEG-2: For higher quality video at a bit-rate of more than 4 Mbps Defined seven profiles aimed at different applications: Simple, Main, SNR scalable, Spatially scalable, High, 4:2:2,

### ITU-T DRAFT H.263 VIDEO CODING FOR LOW BITRATE COMMUNICATION LINE TRANSMISSION OF NON-TELEPHONE SIGNALS. DRAFT ITU-T Recommendation H.

INTERNATIONAL TELECOMMUNICATION UNION ITU-T DRAFT H.263 TELECOMMUNICATION (2 May, 1996) STANDARDIZATION SECTOR OF ITU LINE TRANSMISSION OF NON-TELEPHONE SIGNALS VIDEO CODING FOR LOW BITRATE COMMUNICATION

### Multimedia Communications Multimedia Technologies & Applications. Prof. Dr. Abdulmotaleb El Saddik

Multimedia Communications Multimedia Technologies & Applications Prof. Dr. Abdulmotaleb El Saddik Multimedia Communications Research Laboratory School of Information Technology and Engineering University

### Multimedia Coding and Transmission. Video Coding. Ifi, UiO Norsk Regnesentral Vårsemester 2003 Wolfgang Leister. This part of the course...

Multimedia Coding and Transmission Video Coding Ifi, UiO Norsk Regnesentral Vårsemester 2003 Wolfgang Leister This part of the course...... is held at Ifi, UiO... (Wolfgang Leister) and at University College

### A real-time SNR scalable transcoder for MPEG-2 video streams

EINDHOVEN UNIVERSITY OF TECHNOLOGY Department of Mathematics and Computer Science A real-time SNR scalable transcoder for MPEG-2 video streams by Mohammad Al-khrayshah Supervisors: Prof. J.J. Lukkien Eindhoven

### ADAPTIVE JOINT H.263-CHANNEL CODING FOR MEMORYLESS BINARY CHANNELS

ADAPTIVE JOINT H.263-CHANNEL ING FOR MEMORYLESS BINARY CHANNELS A. Navarro, J. Tavares Aveiro University - Telecommunications Institute, 38 Aveiro, Portugal, navarro@av.it.pt Abstract - The main purpose

### Video Coding in H.26L

Royal Institute of Technology MASTER OF SCIENCE THESIS Video Coding in H.26L by Kristofer Dovstam April 2000 Work done at Ericsson Radio Systems AB, Kista, Sweden, Ericsson Research, Department of Audio

### Module 7 VIDEO CODING AND MOTION ESTIMATION

Module 7 VIDEO CODING AND MOTION ESTIMATION Lesson 20 Basic Building Blocks & Temporal Redundancy Instructional Objectives At the end of this lesson, the students should be able to: 1. Name at least five

### CS 335 Graphics and Multimedia. Image Compression

CS 335 Graphics and Multimedia Image Compression CCITT Image Storage and Compression Group 3: Huffman-type encoding for binary (bilevel) data: FAX Group 4: Entropy encoding without error checks of group

### COMPARATIVE ANALYSIS OF DIRAC PRO-VC-2, H.264 AVC AND AVS CHINA-P7

COMPARATIVE ANALYSIS OF DIRAC PRO-VC-2, H.264 AVC AND AVS CHINA-P7 A Thesis Submitted to the College of Graduate Studies and Research In Partial Fulfillment of the Requirements For the Degree of Master

### EE Low Complexity H.264 encoder for mobile applications

EE 5359 Low Complexity H.264 encoder for mobile applications Thejaswini Purushotham Student I.D.: 1000-616 811 Date: February 18,2010 Objective The objective of the project is to implement a low-complexity

### ADAPTIVE PICTURE SLICING FOR DISTORTION-BASED CLASSIFICATION OF VIDEO PACKETS

ADAPTIVE PICTURE SLICING FOR DISTORTION-BASED CLASSIFICATION OF VIDEO PACKETS E. Masala, D. Quaglia, J.C. De Martin Λ Dipartimento di Automatica e Informatica/ Λ IRITI-CNR Politecnico di Torino, Italy

### Introduction to Video Coding

Introduction to Video Coding o Motivation & Fundamentals o Principles of Video Coding o Coding Standards Special Thanks to Hans L. Cycon from FHTW Berlin for providing first-hand knowledge and much of

### About MPEG Compression. More About Long-GOP Video

About MPEG Compression HD video requires significantly more data than SD video. A single HD video frame can require up to six times more data than an SD frame. To record such large images with such a low

### Multimedia Decoder Using the Nios II Processor

Multimedia Decoder Using the Nios II Processor Third Prize Multimedia Decoder Using the Nios II Processor Institution: Participants: Instructor: Indian Institute of Science Mythri Alle, Naresh K. V., Svatantra

### CMPT 365 Multimedia Systems. Media Compression - Video

CMPT 365 Multimedia Systems Media Compression - Video Spring 2017 Edited from slides by Dr. Jiangchuan Liu CMPT365 Multimedia Systems 1 Introduction What s video? a time-ordered sequence of frames, i.e.,

### Figure 1. Generic Encoder. Window. Spectral Analysis. Psychoacoustic Model. Quantize. Pack Data into Frames. Additional Coding.

Introduction to Digital Audio Compression B. Cavagnolo and J. Bier Berkeley Design Technology, Inc. 2107 Dwight Way, Second Floor Berkeley, CA 94704 (510) 665-1600 info@bdti.com http://www.bdti.com INTRODUCTION

### IMPROVED CONTEXT-ADAPTIVE ARITHMETIC CODING IN H.264/AVC

17th European Signal Processing Conference (EUSIPCO 2009) Glasgow, Scotland, August 24-28, 2009 IMPROVED CONTEXT-ADAPTIVE ARITHMETIC CODING IN H.264/AVC Damian Karwowski, Marek Domański Poznań University

### VIDEO COMPRESSION. Image Compression. Multimedia File Formats. Lossy Compression. Multimedia File Formats. October 8, 2009

File Formats Lossy Compression Image Compression File Formats VIDEO COMPRESSION 121 (Basics) video := time sequence of single images frequent point of view: video compression = image compression with a

### Robert Matthew Buckley. Nova Southeastern University. Dr. Laszlo. MCIS625 On Line. Module 2 Graphics File Format Essay

1 Robert Matthew Buckley Nova Southeastern University Dr. Laszlo MCIS625 On Line Module 2 Graphics File Format Essay 2 JPEG COMPRESSION METHOD Joint Photographic Experts Group (JPEG) is the most commonly

### OVERVIEW OF IEEE 1857 VIDEO CODING STANDARD

OVERVIEW OF IEEE 1857 VIDEO CODING STANDARD Siwei Ma, Shiqi Wang, Wen Gao {swma,sqwang, wgao}@pku.edu.cn Institute of Digital Media, Peking University ABSTRACT IEEE 1857 is a multi-part standard for multimedia

### How to achieve low latency audio/video streaming over IP network?

February 2018 How to achieve low latency audio/video streaming over IP network? Jean-Marie Cloquet, Video Division Director, Silex Inside Gregory Baudet, Marketing Manager, Silex Inside Standard audio

### Megapixel Video for. Part 2 of 4. Brought to You by. Presented by Video Security Consultants

rought to You by 2009 Video Security Consultants Presented by Part 2 of 4 A1 Part 2 of 4 How to Avert a Compression Depression Illustration by Jerry King While bandwidth is widening, larger video systems

### SCALABLE HYBRID VIDEO CODERS WITH DOUBLE MOTION COMPENSATION

SCALABLE HYBRID VIDEO CODERS WITH DOUBLE MOTION COMPENSATION Marek Domański, Łukasz Błaszak, Sławomir Maćkowiak, Adam Łuczak Poznań University of Technology, Institute of Electronics and Telecommunications,

### Fast Implementation of VC-1 with Modified Motion Estimation and Adaptive Block Transform

Circuits and Systems, 2010, 1, 12-17 doi:10.4236/cs.2010.11003 Published Online July 2010 (http://www.scirp.org/journal/cs) Fast Implementation of VC-1 with Modified Motion Estimation and Adaptive Block

### Modeling of an MPEG Audio Layer-3 Encoder in Ptolemy

Modeling of an MPEG Audio Layer-3 Encoder in Ptolemy Patrick Brown EE382C Embedded Software Systems May 10, 2000 \$EVWUDFW MPEG Audio Layer-3 is a standard for the compression of high-quality digital audio.

### Development and optimization of coding algorithms for mobile 3DTV. Gerhard Tech Heribert Brust Karsten Müller Anil Aksay Done Bugdayci

Development and optimization of coding algorithms for mobile 3DTV Gerhard Tech Heribert Brust Karsten Müller Anil Aksay Done Bugdayci Project No. 216503 Development and optimization of coding algorithms

### H.261. Lecture Special Topics in Signal Processing. Multimedia Communications: Coding, Systems, and Networking

8-899 Special Topics in Signal Processing Multimedia Communications: Coding, Sstems, and Networking Prof. Tsuhan Chen tsuhan@ece.cmu.edu Lecture 4 H.6 H.6 ITU-T Stud Group 5, 984-99 Videophone and video

### Lecture 5: Video Compression Standards (Part2) Tutorial 3 : Introduction to Histogram

Lecture 5: Video Compression Standards (Part) Tutorial 3 : Dr. Jian Zhang Conjoint Associate Professor NICTA & CSE UNSW COMP9519 Multimedia Systems S 006 jzhang@cse.unsw.edu.au Introduction to Histogram

### MPEG-4 aacplus - Audio coding for today s digital media world

MPEG-4 aacplus - Audio coding for today s digital media world Whitepaper by: Gerald Moser, Coding Technologies November 2005-1 - 1. Introduction Delivering high quality digital broadcast content to consumers

### Introduction to Video Encoding

Introduction to Video Encoding Preben N. Olsen University of Oslo and Simula Research Laboratory preben@simula.no August 26, 2013 1 / 37 Agenda 1 Introduction Repetition History Quality Assessment Containers

### Parametric Coding of High-Quality Audio

Parametric Coding of High-Quality Audio Prof. Dr. Gerald Schuller Fraunhofer IDMT & Ilmenau Technical University Ilmenau, Germany 1 Waveform vs Parametric Waveform Filter-bank approach Mainly exploits

### Compression and File Formats

Compression and File Formats 1 Compressing Moving Images Methods: Motion JPEG, Cinepak, Indeo, MPEG Known as CODECs compression / decompression algorithms hardware and software implementations symmetrical

### Lecture 7, Video Coding, Motion Compensation Accuracy

Lecture 7, Video Coding, Motion Compensation Accuracy Last time we saw several methods to obtain a good motion estimation, with reduced complexity (efficient search), and with the possibility of sub-pixel

### 06/12/2017. Image compression. Image compression. Image compression. Image compression. Coding redundancy: image 1 has four gray levels

Theoretical size of a file representing a 5k x 4k colour photograph: 5000 x 4000 x 3 = 60 MB 1 min of UHD tv movie: 3840 x 2160 x 3 x 24 x 60 = 36 GB 1. Exploit coding redundancy 2. Exploit spatial and

### ERROR-ROBUST INTER/INTRA MACROBLOCK MODE SELECTION USING ISOLATED REGIONS

ERROR-ROBUST INTER/INTRA MACROBLOCK MODE SELECTION USING ISOLATED REGIONS Ye-Kui Wang 1, Miska M. Hannuksela 2 and Moncef Gabbouj 3 1 Tampere International Center for Signal Processing (TICSP), Tampere,

### High Efficiency Video Coding: The Next Gen Codec. Matthew Goldman Senior Vice President TV Compression Technology Ericsson

High Efficiency Video Coding: The Next Gen Codec Matthew Goldman Senior Vice President TV Compression Technology Ericsson High Efficiency Video Coding Compression Bitrate Targets Bitrate MPEG-2 VIDEO 1994

### Scalable Extension of HEVC 한종기

Scalable Extension of HEVC 한종기 Contents 0. Overview for Scalable Extension of HEVC 1. Requirements and Test Points 2. Coding Gain/Efficiency 3. Complexity 4. System Level Considerations 5. Related Contributions

### IO [io] MAYAH. IO [io] Audio Video Codec Systems

IO [io] MAYAH IO [io] Audio Video Codec Systems MPEG 4 Audio Video Embedded 24/7 Real-Time Solution MPEG 4 Audio Video Production and Streaming Solution ISMA compliant 24/7 Audio Video Realtime Solution

### Scalable Multi-DM642-based MPEG-2 to H.264 Transcoder. Arvind Raman, Sriram Sethuraman Ittiam Systems (Pvt.) Ltd. Bangalore, India

Scalable Multi-DM642-based MPEG-2 to H.264 Transcoder Arvind Raman, Sriram Sethuraman Ittiam Systems (Pvt.) Ltd. Bangalore, India Outline of Presentation MPEG-2 to H.264 Transcoding Need for a multiprocessor

### Anatomy of a Video Codec

Anatomy of a Video Codec The inner workings of Ogg Theora Dr. Timothy B. Terriberry Outline Introduction Video Structure Motion Compensation The DCT Transform Quantization and Coding The Loop Filter Conclusion

### IMAGE COMPRESSION. Chapter - 5 : (Basic)

Chapter - 5 : IMAGE COMPRESSION (Basic) Q() Explain the different types of redundncies that exists in image.? (8M May6 Comp) [8M, MAY 7, ETRX] A common characteristic of most images is that the neighboring

### TKT-2431 SoC design. Introduction to exercises. SoC design / September 10

TKT-2431 SoC design Introduction to exercises Assistants: Exercises and the project work Juha Arvio juha.arvio@tut.fi, Otto Esko otto.esko@tut.fi In the project work, a simplified H.263 video encoder is

### Research on Transcoding of MPEG-2/H.264 Video Compression

Research on Transcoding of MPEG-2/H.264 Video Compression WEI, Xianghui Graduate School of Information, Production and Systems Waseda University February 2009 Abstract Video transcoding performs one or

### MPEG-1. Overview of MPEG-1 1 Standard. Introduction to perceptual and entropy codings

MPEG-1 Overview of MPEG-1 1 Standard Introduction to perceptual and entropy codings Contents History Psychoacoustics and perceptual coding Entropy coding MPEG-1 Layer I/II Layer III (MP3) Comparison and

### For layered video encoding, video sequence is encoded into a base layer bitstream and one (or more) enhancement layer bit-stream(s).

3rd International Conference on Multimedia Technology(ICMT 2013) Video Standard Compliant Layered P2P Streaming Man Yau Chiu 1, Kangheng Wu 1, Zhibin Lei 1 and Dah Ming Chiu 2 Abstract. Peer-to-peer (P2P)

### 2.2: Images and Graphics Digital image representation Image formats and color models JPEG, JPEG2000 Image synthesis and graphics systems

Chapter 2: Representation of Multimedia Data Audio Technology Images and Graphics Video Technology Chapter 3: Multimedia Systems Communication Aspects and Services Chapter 4: Multimedia Systems Storage

### Digital Image Representation. Image Representation. Color Models

Digital Representation Chapter : Representation of Multimedia Data Audio Technology s and Graphics Video Technology Chapter 3: Multimedia Systems Communication Aspects and Services Chapter 4: Multimedia

### 7.5 Dictionary-based Coding

7.5 Dictionary-based Coding LZW uses fixed-length code words to represent variable-length strings of symbols/characters that commonly occur together, e.g., words in English text LZW encoder and decoder

### DIGITAL IMAGE PROCESSING WRITTEN REPORT ADAPTIVE IMAGE COMPRESSION TECHNIQUES FOR WIRELESS MULTIMEDIA APPLICATIONS

DIGITAL IMAGE PROCESSING WRITTEN REPORT ADAPTIVE IMAGE COMPRESSION TECHNIQUES FOR WIRELESS MULTIMEDIA APPLICATIONS SUBMITTED BY: NAVEEN MATHEW FRANCIS #105249595 INTRODUCTION The advent of new technologies

### A Hybrid Temporal-SNR Fine-Granular Scalability for Internet Video

318 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 11, NO. 3, MARCH 2001 A Hybrid Temporal-SNR Fine-Granular Scalability for Internet Video Mihaela van der Schaar, Member, IEEE, and

### Video compression with 1-D directional transforms in H.264/AVC

Video compression with 1-D directional transforms in H.264/AVC The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation Kamisli, Fatih,

### Compressed Audio Demystified by Hendrik Gideonse and Connor Smith. All Rights Reserved.

Compressed Audio Demystified Why Music Producers Need to Care About Compressed Audio Files Download Sales Up CD Sales Down High-Definition hasn t caught on yet Consumers don t seem to care about high fidelity

### Compression; Error detection & correction

Compression; Error detection & correction compression: squeeze out redundancy to use less memory or use less network bandwidth encode the same information in fewer bits some bits carry no information some

### Video Codec Design Developing Image and Video Compression Systems

Video Codec Design Developing Image and Video Compression Systems Iain E. G. Richardson The Robert Gordon University, Aberdeen, UK JOHN WILEY & SONS, LTD Contents 1 Introduction l 1.1 Image and Video Compression

### Ch 4: Multimedia. Fig.4.1 Internet Audio/Video

Ch 4: Multimedia Recent advances in technology have changed our use of audio and video. In the past, we listened to an audio broadcast through a radio and watched a video program broadcast through a TV.

### The MPEG-4 General Audio Coder

The MPEG-4 General Audio Coder Bernhard Grill Fraunhofer Institute for Integrated Circuits (IIS) grl 6/98 page 1 Outline MPEG-2 Advanced Audio Coding (AAC) MPEG-4 Extensions: Perceptual Noise Substitution

### An Improved H.26L Coder Using Lagrangian Coder Control. Summary

UIT - Secteur de la normalisation des télécommunications ITU - Telecommunication Standardization Sector UIT - Sector de Normalización de las Telecomunicaciones Study Period 2001-2004 Commission d' études

### Multimedia Communications. Transform Coding

Multimedia Communications Transform Coding Transform coding Transform coding: source output is transformed into components that are coded according to their characteristics If a sequence of inputs is transformed

### DSP. Presented to the IEEE Central Texas Consultants Network by Sergio Liberman

DSP The Technology Presented to the IEEE Central Texas Consultants Network by Sergio Liberman Abstract The multimedia products that we enjoy today share a common technology backbone: Digital Signal Processing

### Overview: motion-compensated coding

Overview: motion-compensated coding Motion-compensated prediction Motion-compensated hybrid coding Motion estimation by block-matching Motion estimation with sub-pixel accuracy Power spectral density of

### ISO/IEC INTERNATIONAL STANDARD. Information technology Generic coding of moving pictures and associated audio information Part 2: Video

INTERNATIONAL STANDARD ISO/IEC 13818-2 Third edition 2013-10-01 Information technology Generic coding of moving pictures and associated audio information Part 2: Video Technologies de l'information Codage

### MPEG-4. Today we'll talk about...

INF5081 Multimedia Coding and Applications Vårsemester 2007, Ifi, UiO MPEG-4 Wolfgang Leister Knut Holmqvist Today we'll talk about... MPEG-4 / ISO/IEC 14496...... is more than a new audio-/video-codec...

### JPEG. Wikipedia: Felis_silvestris_silvestris.jpg, Michael Gäbler CC BY 3.0

JPEG Wikipedia: Felis_silvestris_silvestris.jpg, Michael Gäbler CC BY 3.0 DFT vs. DCT Image Compression Image compression system Input Image MAPPER QUANTIZER SYMBOL ENCODER Compressed output Image Compression

### A Video CoDec Based on the TMS320C6X DSP José Brito, Leonel Sousa EST IPCB / INESC Av. Do Empresário Castelo Branco Portugal

A Video CoDec Based on the TMS320C6X DSP José Brito, Leonel Sousa EST IPCB / INESC Av. Do Empresário Castelo Branco Portugal jbrito@est.ipcb.pt IST / INESC Rua Alves Redol, Nº 9 1000 029 Lisboa Portugal

### EE482: Digital Signal Processing Applications

Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu EE482: Digital Signal Processing Applications Spring 2014 TTh 14:30-15:45 CBC C222 Lecture 13 Audio Signal Processing 14/04/01 http://www.ee.unlv.edu/~b1morris/ee482/

### Lecture 16 Perceptual Audio Coding

EECS 225D Audio Signal Processing in Humans and Machines Lecture 16 Perceptual Audio Coding 2012-3-14 Professor Nelson Morgan today s lecture by John Lazzaro www.icsi.berkeley.edu/eecs225d/spr12/ Hero

### MULTIPLEXING/DE-MULTIPLEXING DIRAC VIDEO WITH AAC AUDIO BIT STREAM ASHWINI URS. Presented to the Faculty of the Graduate School of

MULTIPLEXING/DE-MULTIPLEXING DIRAC VIDEO WITH AAC AUDIO BIT STREAM by ASHWINI URS Presented to the Faculty of the Graduate School of The University of Texas at Arlington in Partial Fulfillment of the Requirements

### R 118 TIERING OF CAMERAS FOR USE IN TELEVISION PRODUCTION

R 118 TIERING OF CAMERAS FOR USE IN TELEVISION PRODUCTION VERSION 2 Geneva April 2017 Page intentionally left blank. This document is paginated for two sided printing EBU R 118 Tiering of Cameras for use

### The Best-Performance Digital Video Recorder JPEG2000 DVR V.S M-PEG & MPEG4(H.264)

The Best-Performance Digital Video Recorder JPEG2000 DVR V.S M-PEG & MPEG4(H.264) Many DVRs in the market But it takes brains to make the best product JPEG2000 The best picture quality in playback. Brief

### FPGA IMPLEMENTATION OF BIT PLANE ENTROPY ENCODER FOR 3 D DWT BASED VIDEO COMPRESSION

FPGA IMPLEMENTATION OF BIT PLANE ENTROPY ENCODER FOR 3 D DWT BASED VIDEO COMPRESSION 1 GOPIKA G NAIR, 2 SABI S. 1 M. Tech. Scholar (Embedded Systems), ECE department, SBCE, Pattoor, Kerala, India, Email: