DSP. Presented to the IEEE Central Texas Consultants Network by Sergio Liberman

Similar documents
Audio-coding standards

Audio-coding standards

Both LPC and CELP are used primarily for telephony applications and hence the compression of a speech signal.

Perceptual coding. A psychoacoustic model is used to identify those signals that are influenced by both these effects.

Introducing Audio Signal Processing & Audio Coding. Dr Michael Mason Snr Staff Eng., Team Lead (Applied Research) Dolby Australia Pty Ltd

Introducing Audio Signal Processing & Audio Coding. Dr Michael Mason Senior Manager, CE Technology Dolby Australia Pty Ltd

Audio Compression. Audio Compression. Absolute Threshold. CD quality audio:

Chapter 14 MPEG Audio Compression

Appendix 4. Audio coding algorithms

Mpeg 1 layer 3 (mp3) general overview

Multimedia Communications. Audio coding

EE482: Digital Signal Processing Applications

Perceptual Coding. Lossless vs. lossy compression Perceptual models Selecting info to eliminate Quantization and entropy encoding

Fundamentals of Perceptual Audio Encoding. Craig Lewiston HST.723 Lab II 3/23/06

Audio Fundamentals, Compression Techniques & Standards. Hamid R. Rabiee Mostafa Salehi, Fatemeh Dabiran, Hoda Ayatollahi Spring 2011

Lecture 16 Perceptual Audio Coding

CHAPTER 10: SOUND AND VIDEO EDITING

Figure 1. Generic Encoder. Window. Spectral Analysis. Psychoacoustic Model. Quantize. Pack Data into Frames. Additional Coding.

Principles of Audio Coding

2.4 Audio Compression

The MPEG-4 General Audio Coder

Audio Coding and MP3

Optical Storage Technology. MPEG Data Compression

Squeeze Play: The State of Ady0 Cmprshn. Scott Selfon Senior Development Lead Xbox Advanced Technology Group Microsoft

5: Music Compression. Music Coding. Mark Handley

CSCD 443/533 Advanced Networks Fall 2017

Compressed Audio Demystified by Hendrik Gideonse and Connor Smith. All Rights Reserved.

ELL 788 Computational Perception & Cognition July November 2015

Chapter 5.5 Audio Programming

Implementation of a MPEG 1 Layer I Audio Decoder with Variable Bit Lengths

Networking Applications

Parametric Coding of High-Quality Audio

Contents. 3 Vector Quantization The VQ Advantage Formulation Optimality Conditions... 48

MPEG-4 General Audio Coding

Digital Media. Daniel Fuller ITEC 2110

Data Compression. Audio compression

DAB. Digital Audio Broadcasting

ITNP80: Multimedia! Sound-II!

Module 9 AUDIO CODING. Version 2 ECE IIT, Kharagpur

MPEG-1. Overview of MPEG-1 1 Standard. Introduction to perceptual and entropy codings

Parametric Coding of Spatial Audio

Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig

CISC 7610 Lecture 3 Multimedia data and data formats

Audio and video compression

Classification of Semiconductor LSI

Lossy compression. CSCI 470: Web Science Keith Vertanen

Principles of MPEG audio compression

ITEC310 Computer Networks II

KINGS COLLEGE OF ENGINEERING DEPARTMENT OF INFORMATION TECHNOLOGY ACADEMIC YEAR / ODD SEMESTER QUESTION BANK

For Mac and iphone. James McCartney Core Audio Engineer. Eric Allamanche Core Audio Engineer

Efficient Implementation of Transform Based Audio Coders using SIMD Paradigm and Multifunction Computations

Digital Recording and Playback

Modeling of an MPEG Audio Layer-3 Encoder in Ptolemy

REAL TIME DIGITAL SIGNAL PROCESSING

Rich Recording Technology Technical overall description

Digital Signal Processor 2010/1/4

CS 074 The Digital World. Digital Audio

Ch. 5: Audio Compression Multimedia Systems

Evaluating MMX Technology Using DSP and Multimedia Applications

Lecture #3: Digital Music and Sound

Filterbanks and transforms

2008/12/23. System Arch 2008 (Fire Tom Wada) 1

DuSLIC Infineons High Modem Performance Codec

Efficient Representation of Sound Images: Recent Developments in Parametric Coding of Spatial Audio

Chapter 4: Audio Coding

The following bit rates are recommended for broadcast contribution employing the most commonly used audio coding schemes:

AUDIOVISUAL COMMUNICATION

New Results in Low Bit Rate Speech Coding and Bandwidth Extension

Audio Processing on ARM Cortex -M4 for Automotive Applications

Multimedia Systems Speech I Mahdi Amiri February 2011 Sharif University of Technology

3 Sound / Audio. CS 5513 Multimedia Systems Spring 2009 LECTURE. Imran Ihsan Principal Design Consultant

Multimedia Systems Speech I Mahdi Amiri September 2015 Sharif University of Technology

Computing in the Modern World

FILE CONVERSION AFTERMATH: ANALYSIS OF AUDIO FILE STRUCTURE FORMAT

1 Audio quality determination based on perceptual measurement techniques 1 John G. Beerends

Video Compression An Introduction

2014 Summer School on MPEG/VCEG Video. Video Coding Concept

MPEG-4 aacplus - Audio coding for today s digital media world

Embedding Audio into your RX Application

UNDERSTANDING MUSIC & VIDEO FORMATS

Port of a Fixed Point MPEG-2 AAC Encoder on a ARM Platform

Image and Video Coding I: Fundamentals

Audio Coding Standards

REAL TIME DIGITAL SIGNAL PROCESSING

Fundamental of Digital Media Design. Introduction to Audio

University of Saskatchewan 5-1 EE 392 Electrical Engineering Laboratory III

Lossy compression CSCI 470: Web Science Keith Vertanen Copyright 2013

AUDIO MEDIA CHAPTER Background

CHAPTER 2 - DIGITAL DATA REPRESENTATION AND NUMBERING SYSTEMS

Lecture Information Multimedia Video Coding & Architectures

Enhanced Audio Features for High- Definition Broadcasts and Discs. Roland Vlaicu Dolby Laboratories, Inc.

S.K.R Engineering College, Chennai, India. 1 2

ABSTRACT. that it avoids the tolls charged by ordinary telephone service

A Digital Audio Primer

SPREAD SPECTRUM AUDIO WATERMARKING SCHEME BASED ON PSYCHOACOUSTIC MODEL

D. Richard Brown III Professor Worcester Polytechnic Institute Electrical and Computer Engineering Department

Storage I/O Summary. Lecture 16: Multimedia and DSP Architectures

THE FUTURE OF LISTENING

What is multimedia? Multimedia. Continuous media. Most common media types. Continuous media processing. Interactivity. What is multimedia?

AUDIOVISUAL COMMUNICATION

Transcription:

DSP The Technology Presented to the IEEE Central Texas Consultants Network by Sergio Liberman

Abstract The multimedia products that we enjoy today share a common technology backbone: Digital Signal Processing (DSP). The digital revolution for consumers started with the introduction of the Compact Disc (CD) format for music distribution. With the increasing computing power and lower cost of digital processing resulting from Moore's Law, advanced signal processing algorithms have allowed consumer products such as HDTV, DVD and Blu-ray, MP3 players, multimedia capable smart phones, and audio/video streaming. This presentation will look into how analog signals such as sound and pictures are brought in and out of the digital domain for communication, storage, transmission, processing and playback, as well as some of the DSP algorithms. 2

Signals Mathematically, a signal is described as a function of one or more independent variables Signals in the physical world are any time varying or spatial varying quantity Signals in electronic systems are in the form of varying electrical voltage or current Signal variations carry information Examples of signals: Sound signals: f(t) => variation of air pressure in time Image signals: f(x,y) => variation of intensity in space Video signals: f(x,y,t) => variation of intensity in space & time 3

Signal Processing Modification of signals by processing elements. Linear processing: amplification, filtering, tone controls, equalization, etc. Non-linear processing: dynamic range control, noise gating, etc. Spectral analysis, synthesis, signal generation, etc. 4

Analog Signal Processing Analog signal processing is any signal processing conducted on analog signals by analog transducers and analog processing elements. Electrical analog processing elements include capacitors, resistors, inductors, transistors and operational amplifiers. Analog transducers include microphones, speakers, record player and guitar pickups, magnetic tape heads, etc. 5

Analog Signal Processing Disadvantages The primary disadvantage of analog signal processing is that any system has noise i.e., random unwanted variation. As the signal is copied and re-copied, or transmitted over long distances, these apparently random variations become dominant. Analog processing components suffer from aging, drift with temperature, tuning issues, accuracy, repeatability, size, etc. 6

Digital Signal Processing Advantages Representation of signals by a sequence of numbers or symbols, which can be copied and transmitted with no noise or errors in a digital system. Digital processing components and their numerical results do not suffer from aging, drift with temperature, tuning issues, accuracy, repeatability, etc. There are signal processing functions only possible with DSP, such as linear phase filters. Moore s Law allows for huge size reduction of digital processing elements. Combination of advanced control and data processing with signal processing. 7

Digital Signal Processing Use digital computer processing elements for manipulation of signals. Digital processing elements cannot handle infinite values for the signal variables and parameters. Requires analog signals to be converted from continuous time/space to discrete time/space. Requires quantizing continuous analog signal amplitude values to a finite set of discrete signal amplitude values. 8

Nyquist Shannon Sampling Theorem A band-limited analog signal, where Bis its highest frequency, can be perfectly reconstructed from an infinite sequence of its samples if the sampling rate exceeds 2B samples per second. Allows sampling of analog signal, changing continuous time/space to a series of discrete signal value points. 9

Quantization of Amplitude After sampling of the analog signal, the amplitude values need to be quantized to a finite range. Quantization introduces amplitude error, resulting in quantization noise. Increasing resolution (the number of possible values for the amplitude) reduces quantization noise. Proper digital signal processing assumes enough bits of resolution in the A/D process so that quantization noise is negligible for the application. 10

How Many Bits Are Necessary? Every additional bit in resolution adds ~6dB to the SNR (signal to noise ratio) for uniform quantization. The dynamic range of human hearing is roughly 120 db. (20 bits) The dynamic range of music as normally perceived in a concert hall doesn't exceed 80 db (14 bits) Human speech is normally perceived over a range of about 40 db. (7 bits) 11

ADC: Analog to Digital Conversion Converts the analog signal into a digital representation by sampling and amplitude quantization. 12

DAC: Digital to Analog Conversion DAC converts the digital signal abstract numbers into a concrete sequence of impulses that are then processed by a reconstruction filter using some form of interpolation to fill in data between the impulses. Delta-sigma modulation method generates a pulsedensity modulated signal that can then be filtered to produce a smoothly varying analog signal. Most popular DAC for audio. PWM modulation method generates a pulse-width modulated signal that can then be filtered. Used in Class-D amplifers. 13

Example: Compact Disc digital audio First digital consumer format -1980 The CD disc contains audio signals converted from analog to digital using a linear encoding format. Audio bandwidth is assumed to be limited to 20 khz (B from Nyquist-Shannon). The sampling rate for CD audio is 44,100 samples per second ( > 2 x 20 khz). The amplitude values are quantized to a fixed-point binary representation using 16 bits per audio sample providing 96dB SNR. The audio data bit rate for a CD disc is 16 bits x 44,100 samples/s x 2 channels = 1.4 Mbit/s.WAV file extension. 14

Example: T-Carrier Digital Telephony First extensive use of DSP -1961 Telephony audio bandwidth is 3.5 khz (B). Voice in digital telephony is sampled at 8,000 samples per second (> 2 x 3.5kHz). Amplitude of voice signals are quantized to ~ 13.5 bits per sample (equivalent linear encoding). The amplitude is compressed using A-law/u-law non-linear compandingalgorithm to reduce the number of bits per sample to 8 bits. Bit rate for one voice circuit: 8 bits x 8,000 samples/s = 64 kbit/s 15

Signal Compression -1961 A-law/u-law non-linear compandingreduces the dynamic range of the signal, thereby increasing the coding efficiency and resulting in a signal-todistortion ratio that is superior to that obtained by linear encoding for a given number of bits. Bit rate for one voice circuit: 8 bits x 8,000 samples/s = 64 kbit/s Bit rate before companding: 14 bits x 8,000 samples/s = 112 kbit/s 16

CompandingTransfer Function 17

Coding Coding, in the sense used here, is the process of reducing the bit rate of a digital signal. Reduces storage requirements and transmission bandwidth for the digital signal. The coder input is a digital signal. The coder output is a lower bit rate digital signal. The decoder reverses the process and provides (an approximation to) the original digital signal. 18

Types of Coding Lossless Coding commonly refers to coding methods that are completely reversible, where the original signal can be reconstructed bit for bit. Equivalent to data compression in computers such as Zip, RAR, etc. Lossycoding commonly refers to coders that create an approximate reproduction of their input signal. The nature of the loss depends entirely on the kind of lossycoding used. MP3 is an example of lossy coding. 19

Perceptual Coding Perceptual coding uses a model of the destination, i.e. the human being who will be using the data. Perceptual coding attempts to remove parts of the signal that the human cannot perceive. Perceptual coding is lossy. Perceptual coders will, in general, have a lower SNR than a source coder, and a higher perceived quality than a source coder of equivalent bit rate. 20

Using Perceptual Coding Perceptual audio coding is applicable where the signal will NOT be reprocessed, equalized, or otherwise modified before the final delivery to the consumer. NEVER use multiple passes through the perceptual encoders. Popular audio perceptual coding: MP3, AAC, Dolby Digital, DTS 21

Perceptual Audio Coder Block Diagram Filtered Audio Signal Audio Input Coding Filterbank Quantization and Rate Control Noiseless Coding and Bitstream Packing Perceptual Model Perceptual Threshold Quantized filterbank values, side information Coded Bitstream 22

What is Auditory Masking? The Human Auditory System (HAS) has a limited detection ability when a stronger signal occurs near (in frequency and time) to a weaker signal. In many situations, the weaker signal is imperceptible even under ideal listening conditions. 23

Auditory Masking Noise Masker Tone Probe The Tone is NOT Audible 24

Auditory Filterbank The mechanism in the human cochlea constitute a mechanical filterbank. The shape of the filter at any one position on the cochlea is called the cochlear filterfor that point on the cochlea. A critical bandis very close to the passband bandwidth of that filter. 25

Critical Band Critical band is a range of frequencies over which the masking SNR remains more or less constant. For example, any noise signal within +-.5 critical band of the tone will produce nearly the same masking behavior as any other, as long as their energies are the same. 26

Time Considerations in Masking Simultaneous Masking Forward Masking Masking of a signal by a masker that precedes the masked signal. Backward Masking Masking of a signal by a masker that comes after the signal. 27

Forward Masking The length of forward masking is >20ms, and is sometimes stated to be as long as several hundred milliseconds. In practice, the decay for post masker masking has two parts, a short hangoverpart and then a longer decayingpart. Different coders take advantage of this in different ways. 28

Backward Masking Backward masking appears to be due to the length of the impulse response of the cochlear filter. At high frequencies, backward masking is less than 1ms for a trained subject who is sensitive to monaural time-domain masking effects. Subjects vary significantly in their ability to detect backwardly masked probes. 29

MP3 MP3 = MPEG-1 or MPEG-2 Audio Layer III Standard released in 1993 De facto standard for digital audio compression in portable players Lossy perceptual coding compression MP3 compression to 128 kbit/s data rate achieves ~11x data rate reduction as compared to CD audio Highest bit rate is 320 kbit/s. Decent audio quality for popular music achieved at about 192 kbit/s 30

Limitations of the MP3 Format Time resolution can be too low for highly transient signals and may cause smearing of percussive sounds. Due to the tree structure of the filter bank, pre-echo problems are made worse, as the combined impulse response of the two filter banks does not, and cannot, provide an optimum solution in time/frequency resolution. The combining of the two filter banks' outputs creates aliasing problems that must be handled partially by the "aliasing compensation" stage; however, that creates excess energy to be coded in the frequency domain, thereby decreasing coding efficiency. Frequency resolution is limited by the small long block window size, which decreases coding efficiency. There is no scale factor band for frequencies above 15.5/15.8 khz. Joint stereo is done only on a frame-to-frame basis. Internal handling of the bit reservoir increases encoding delay. Encoder/decoder overall delay is not defined, which means there is no official provision for gapless playback. However, some encoders such as LAME can attach additional metadata that will allow players that can handle it to deliver seamless playback. The data stream can contain an optional checksum, but the checksum onlyprotects the header data, not the audio data. 31

AAC Advanced Audio Coding, designed to replace MP3, used in the Apple ipod as the standard algorithm. Lossy perceptual coding compression. Achieves better sound quality for the same bit rate as compared to MP3. 32

FIR Filter A discrete-time FIR filter of order N. The top part is an N-stage delay line with N+1 taps. 33

FIR Filter (cont.) A finite impulse response (FIR)filter is a type of a signal processing filter whose impulse response is of finite duration. Only filter capable of linear phase. No feedback, so it is inherently stable, and there is no recirculation of rounding errors. The output is calculated by multiplying each data sample in the delay line by the correspondent filter tap coefficient, and accumulating all multiplies. 34

DSP Processor Features Separate program and data memories, concurrent access on multiple data busses. Special SIMD (single instruction, multiple data) operations. Some processors use VLIW techniques so each instruction drives multiple arithmetic units in parallel. Fast multiply-accumulates (MACs). Many fundamental DSP algorithms, such as FIR Filter or the Fast Fourier Transform (FFT) depend heavily on multiply-accumulate performance. Hardware modulo addressing, allowing circular buffers to be implemented without having to constantly test for wrap around. Bit-reversed addressing, a special data memory addressing mode, useful for calculating FFTs. Special loop controls, such as architectural support for executing a few instruction words in a very tight loop without overhead for instruction fetches or exit testing. Fast interrupt response with low overhead context switching. 35

Freescale DSP56371 Block Diagram 36

Thank You! Sergio Liberman Sergio.Liberman@gmail.com 37