An investigation of non-uniform bandwidths auditory filterbank in audio coding

Similar documents
Lecture 16 Perceptual Audio Coding

Both LPC and CELP are used primarily for telephony applications and hence the compression of a speech signal.

Perceptual coding. A psychoacoustic model is used to identify those signals that are influenced by both these effects.

Audio-coding standards

Audio-coding standards

MPEG-1. Overview of MPEG-1 1 Standard. Introduction to perceptual and entropy codings

Mpeg 1 layer 3 (mp3) general overview

Improved Audio Coding Using a Psychoacoustic Model Based on a Cochlear Filter Bank

Audio Compression Using Decibel chirp Wavelet in Psycho- Acoustic Model

Chapter 14 MPEG Audio Compression

Principles of Audio Coding

EE482: Digital Signal Processing Applications

Efficient Signal Adaptive Perceptual Audio Coding

Perceptual Coding. Lossless vs. lossy compression Perceptual models Selecting info to eliminate Quantization and entropy encoding

Parametric Coding of High-Quality Audio

An adaptive wavelet-based approach for perceptual low bit rate audio coding attending to entropy-type criteria

Multimedia Communications. Audio coding

AUDIOVISUAL COMMUNICATION

Appendix 4. Audio coding algorithms

Wavelet filter bank based wide-band audio coder

A PSYCHOACOUSTIC MODEL WITH PARTIAL SPECTRAL FLATNESS MEASURE FOR TONALITY ESTIMATION

Compressed Audio Demystified by Hendrik Gideonse and Connor Smith. All Rights Reserved.

<< WILL FILL IN THESE SECTIONS THIS WEEK to provide sufficient background>>

Filterbanks and transforms

Optical Storage Technology. MPEG Data Compression

New Results in Low Bit Rate Speech Coding and Bandwidth Extension

Module 9 AUDIO CODING. Version 2 ECE IIT, Kharagpur

AUDIOVISUAL COMMUNICATION

Modeling of an MPEG Audio Layer-3 Encoder in Ptolemy

CT516 Advanced Digital Communications Lecture 7: Speech Encoder

1 Introduction. 2 Speech Compression

Lecture 5: Error Resilience & Scalability

AUDIO COMPRESSION ATLO W BIT RATES USING A SIGNAL ADAPTIVE SWITCHED FILTERBANK. AT&T Bell Labs. Murray Hill, NJ 07974

Pyramid Coding and Subband Coding

Effect of MPEG Audio Compression on HMM-based Speech Synthesis

DSP. Presented to the IEEE Central Texas Consultants Network by Sergio Liberman

Speech and audio coding

MUSIC A Darker Phonetic Audio Coder

Pyramid Coding and Subband Coding

19 th INTERNATIONAL CONGRESS ON ACOUSTICS MADRID, 2-7 SEPTEMBER 2007

Data Compression. Audio compression

5: Music Compression. Music Coding. Mark Handley

Audio Coding and MP3

ELL 788 Computational Perception & Cognition July November 2015

LOW COMPLEXITY SUBBAND ANALYSIS USING QUADRATURE MIRROR FILTERS

Introducing Audio Signal Processing & Audio Coding. Dr Michael Mason Snr Staff Eng., Team Lead (Applied Research) Dolby Australia Pty Ltd

Perceptual Audio Coders What to listen for: Artifacts of Parametric Coding

2.4 Audio Compression

Figure 1. Generic Encoder. Window. Spectral Analysis. Psychoacoustic Model. Quantize. Pack Data into Frames. Additional Coding.

Introducing Audio Signal Processing & Audio Coding. Dr Michael Mason Senior Manager, CE Technology Dolby Australia Pty Ltd

Audio Compression. Audio Compression. Absolute Threshold. CD quality audio:

Contents. 3 Vector Quantization The VQ Advantage Formulation Optimality Conditions... 48

Audio Fundamentals, Compression Techniques & Standards. Hamid R. Rabiee Mostafa Salehi, Fatemeh Dabiran, Hoda Ayatollahi Spring 2011

DSP-CIS. Part-IV : Filter Banks & Subband Systems. Chapter-10 : Filter Bank Preliminaries. Marc Moonen

Scalable Perceptual and Lossless Audio Coding based on MPEG-4 AAC

Speech-Music Discrimination from MPEG-1 Bitstream

Ch. 5: Audio Compression Multimedia Systems

CHAPTER 4 REVERSIBLE IMAGE WATERMARKING USING BIT PLANE CODING AND LIFTING WAVELET TRANSFORM

Efficient Representation of Sound Images: Recent Developments in Parametric Coding of Spatial Audio

Compression transparent low-level description of audio signals

An Ultra Low-Power WOLA Filterbank Implementation in Deep Submicron Technology

Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig

HAVE YOUR CAKE AND HEAR IT TOO: A HUFFMAN CODED, BLOCK SWITCHING, STEREO PERCEPTUAL AUDIO CODER

THE EFFECT OF EARLY REFLECTIONS ON PERCEIVED TIMBRE ANALYZED WITH AN AUDITORY MODEL. Tapio Lokki, Ville Pulkki, and Lauri Savioja

Chapter 2 Studies and Implementation of Subband Coder and Decoder of Speech Signal Using Rayleigh Distribution

Adaptive Quantization for Video Compression in Frequency Domain

Perceptually motivated Sub-band Decomposition for FDLP Audio Coding

ijdsp Interactive Illustrations of Speech/Audio Processing Concepts

SPREAD SPECTRUM AUDIO WATERMARKING SCHEME BASED ON PSYCHOACOUSTIC MODEL

A NEW DCT-BASED WATERMARKING METHOD FOR COPYRIGHT PROTECTION OF DIGITAL AUDIO

Principles of MPEG audio compression

The MPEG-4 General Audio Coder

The following bit rates are recommended for broadcast contribution employing the most commonly used audio coding schemes:

Audio Coding Standards

Simple Watermark for Stereo Audio Signals with Modulated High-Frequency Band Delay

Lecture 7: Audio Compression & Coding

MPEG-1 Bitstreams Processing for Audio Content Analysis

Parametric Coding of Spatial Audio

Fundamentals of Perceptual Audio Encoding. Craig Lewiston HST.723 Lab II 3/23/06

Zhitao Lu and William A. Pearlman. Rensselaer Polytechnic Institute. Abstract

MPEG-4 General Audio Coding

Packet Loss Concealment for Audio Streaming based on the GAPES and MAPES Algorithms

Key words: B- Spline filters, filter banks, sub band coding, Pre processing, Image Averaging IJSER

MPEG-4 aacplus - Audio coding for today s digital media world

DAB. Digital Audio Broadcasting

Perfect Reconstruction FIR Filter Banks and Image Compression

Lecture 12 Video Coding Cascade Transforms H264, Wavelets

Data Hiding in Video

REDUCED FLOATING POINT FOR MPEG1/2 LAYER III DECODING. Mikael Olausson, Andreas Ehliar, Johan Eilert and Dake liu

Bit or Noise Allocation

Audio and video compression

Module 8: Video Coding Basics Lecture 42: Sub-band coding, Second generation coding, 3D coding. The Lecture Contains: Performance Measures

Fatima Michael College of Engineering & Technology

AN AUDIO WATERMARKING SCHEME ROBUST TO MPEG AUDIO COMPRESSION

CISC 7610 Lecture 3 Multimedia data and data formats

14th European Signal Processing Conference (EUSIPCO 2006), Florence, Italy, September 4-8, 2006, copyright by EURASIP

An Experimental High Fidelity Perceptual Audio Coder

Audio Engineering Society. Convention Paper

/ / _ / _ / _ / / / / /_/ _/_/ _/_/ _/_/ _\ / All-American-Advanced-Audio-Codec

Perceptually-Based Joint-Program Audio Coding

Transcription:

PAGE 360 An investigation of non-uniform bandwidths auditory filterbank in audio coding Andrew Lin, Stevan Berber, Waleed Abdulla Department of Electrical and Computer Engineering University of Auckland, New Zealand alin051@ec.auckland.ac.nz, s.berber@auckland.ac.nz, w.abdulla@auckland.ac.nz Abstract This paper presents an investigation on the use of non-linear auditory filterbank in wideband audio coding. The perceptually based parameterization of the audio signal using gammatone filterbank is examined and discussed. Conventional gammatone filters requires high order FIR filters in the synthesis stage which introduces long delay and large computation cost. Here, a simple and efficient synthesis technique is investigated and embedded into a perceptual audio coder. A limitation of this system is that the outputs of the filterbank can not be maximally decimated due to the non-uniform bandwidths of the filterbank, yet the coder achieves near transparent quality at a bit rate of 120 kbps. 1. Introduction Much of the popular audio coders nowadays are perceptual audio coders (PAC), which perform lossy data compression by minimizing the perceived distortion (M. Bosi, 2003; Marina Bosi & Goldberg, 2003; Sinha, Johnston, Dorward, & Quackenbush, 1998). Models of auditory system are employed to determine the masking threshold, which defines the boundary of audible and inaudible parts of the signal. Compression is achieved by reducing the number of bits allocated to the audio signal such that quantization noise is below the masking threshold, hence inaudible. The coding gain is very much depended upon the quality of the psychoacoustic model and the parameterization of the signal (Marina Bosi & Goldberg, 2003). A well-defined psychoacoustic model that closely mimics the auditory system can save bits from signal redundancies, while efficient parameterization of the signal can reduce the signal irrelevancies. The major drawback in the current MPEG-1 audio coders is that it fails to parameterize the signal suitable for the auditory system. In MPEG-1, uniform (equal bandwidth) spectral decomposition is performed to account for the band-pass filtering effect of the cochlea (Baumgarte, 2002). However, the actual auditory system performs non-uniform spectral decomposition (Fletcher, 1940; Moore & Glasberg, 1983). Uniform bandwidth filterbanks were initially used because of their low computational cost. This paper describes the use of auditory-domain filterbanks in wideband audio coding. Gammatone filterbank (GTFB) has been selected because of its inherent auditory-based properties (de Boer, 1975; Patterson et al., 1992; Slaney, 1993). GTFB have been extensively utilized in computational auditory models of the cochlea (Lopez-Poveda & Meddis, 2001; Patterson et al., 1992) and speech processing (Abdulla, 2002; Ambikairajah & Gunawan, 2004; Ambikairajah, Lin, & Holmes, 2004), but it has not been well investigated in audio and speech coding. Kubin (Kubin & Bastiaan Kleijn, 1999) first started the use of gammatone filterbank in speech coding based on the model of auditory neuron firing behavior. Ambikairajah and Lin (Ambikairajah, Epps, & Lin, 2001) further enhanced the coding in (Kubin & Bastiaan Kleijn, 1999) by introducing the temporal and simultaneous masking properties. Both implementations though, have only studied the coding of speech and audio with 8 khz signal bandwidth. In this paper, wideband 44.1 khz sampled audio signals are considered. Furthermore, we investigate a simple and efficient synthesis approach presented in (Hohmann, 2002) instead of the usual timereversed FIR synthesis technique found in (Ambikairajah et al., 2001; Kubin & Bastiaan Kleijn, 1999). This paper is organised as follows: section 2 starts with an overview of the conventional analysis/synthesis GTFB, followed by a discussion on the issues observed when they are used in wideband audio coding. Then, a more efficient implementation of the GTFB from (Hohmann, 2002) is described. Section 3 presents the

employment of such GTFB in a perceptual audio coder and the limitation of this coder is discussed. Section 4 concludes this paper. 2. Gammatone Analysis /Synthesis Filterbank Gammatone filters are a special type of filters whose pass-band bandwidth matches closely to the equivalent rectangular bandwidth (ERB) of the our auditory system (Patterson et al., 1992). The derivation of ERB is based on the theory of critical bandwidth except it provides finer measurement for frequencies below 500 Hz. the analytical expression of ERB as a function of its center frequency fc is defined by: ERB l + f c / q where l = 24.7 and q = 9.265 = (1) The ERB is associated with the gammatone filter by the filter s sampled impulse response: g( n) = A( nt ) e N 1 2πERB( f c ) nt cos(2πf cnt + ϕ) (2) Where A is the gain factor for unity magnitude response, T is the sampling period, N is the order of the filter, fc is the centre frequency of the filter and φ is the phase term for complex filters. Patterson (Patterson et al., 1992) has shown that 4 th order filter (N=4) is a good approximation of the human auditory response. Discrete gammatone filters can be implemented with FIR or IIR scheme. For FIR, the gammatone impulse response in (2) specifies the coefficients of the FIR filter, and for IIR, a 4 th order IIR gammatone filter can be implemented by cascading four first-order band-pass filter (Hohmann, 2002; Slaney, 1993). An all-pole approximation IIR can also be used for simplicity because zeros have a small effect near the centre frequency of each filter (Slaney, 1993). The transfer function of an all-pole 4 th order gammatone filter is defined by: 1 G( z) = (1 ~ (3) αz 1 ) 4 where the coefficients α ~ can be real or complex. A general block diagram of the analysis/synthesis GTFB is shown in Figure 1, which consists of K gammatone filters connected in parallel and each channel is a gammatone band-pass filter centered at f c1 to f ck. The synthesis GTFB is usually implemented using the time-reversed impulse response of the analysis filter, as found in (Ambikairajah et al., 2001; Kubin & Bastiaan Kleijn, 1999). Figure 1: A conventional gammatone analysis and synthesis filterbank Ideally the center frequencies of each channel should be designed such that there is no gap between each pass-band. A so-called ERB scale is defined in (Hohmann, 2002) to measure the relative distance between each filter. The ERB scale is almost logarithmic and is defined by: f ERB scale ( f ) = q.log(1 + ) (4) l. q where l = 24.7 and q = 9.265 By placing the center frequency of each channel at fixed steps on the ERB scales, the pass-bands of the filters cover the whole spectrum without any gap. A fully covered frequency spectrum ranging from 50 to 22050 Hz requires 40 gammatone filers as shown in Figure 2. Figure 2: Frequency responses of 40 individual gammatone filters covering from 50 to 22050 Hz. 2.1 Gammatone filterbank design issues The analysis/synthesis filterbank can be implemented in FIR/FIR, IIR/FIR, or IIR/IIR scheme (Lin, Ambikairajah, & Holmes, 2001). The major problem with the use of FIR is that high order filters are usually required to accurately define the gammatone filters centered at low frequencies. For example, the impulse response of a gammatone filter centered at 100 Hz with a sampling rate of 44.1 khz is shown in Figure 3. To measure the effective length of this impulse response, a PAGE 361

method based on the accumulated energy of the impulse response from (Laakso & Valimaki, 1999) is used. The effective length for 95% accumulated energy of the impulse response in Figure 3 is 1130 samples, hence, the FIR gammatone filters would require at least 1130 taps. Another analysis/synthesis GTFB system proposed in (Hohmann, 2002) avoids the use of FIR filters in synthesis. Each output of the analysis filterbank is delayed by a calculated amount such that the maximum of their real part align at the same time instant. Signal reconstruction is then performed by a weighted summation of the aligned signals in each channel. The block diagram of such analysis/synthesis filterbank system is shown in Figure 5. PAGE 362 Figure 3: Impulse response of a gammatone filter centered at 100 Hz with a sampling rate of 44.1 khz. High order FIR is computationally expensive and the large delay associated with long filters is undesirable, but insufficient length of the filters can result in distortion of the reconstructed signal especially at low frequencies. Figure 4 shows the overall frequency response of a 40 channel FIR/FIR analysis/synthesis filterbank system at low frequencies. The system bandwidth is ranging from 100 to 22050 Hz and the filter length is set to be 50% accumulated energy of the impulse response of the lowest frequency channel. Figure 5: Gammatone analysis/synthesis filterbank with a constant delay. Given a desired overall system delay, the number of samples to be delayed for a particular channel is calculated by the relative position between the impulse response envelope peak and the desired delay position. Figure 6(a) shows the impulse response of channel 20 gammatone filter to be delayed to the 4ms position, that is, sample number 177 with a sampling rate of 44.1kHz shown by the vertical line. Figure 6(b) illustrates the delayed impulse response. Figure 4: Overall analysis/synthesis filterbank frequency response at low frequencies. Insufficient filter length causes large ripples at low frequencies. (a) (b) Subjective listening tests suggest that the filter length should be kept at least 80% accumulated energy of the impulse response of the lowest frequency channel to avoid distortion, this usually requires the FIR filters to have more than 1000 coefficients. 2.2 Gammatone Filterbank with a constant delay Figure 6: (a) The envelope peak of the impulse response is to be delayed to the desire delay position (vertical line). (b) The delayed impulse response. For lower frequency channels the position of the envelope peak may be greater than that of the desired delay, as shown in Figure 7(a). In that case, the first local maximum of the real part of that impulse response is delayed to the desire delay position, depicted in Figure 7(b).

computational cost and delay from the high order FIR filters are avoided. However, the optimal delay factor for the system needs to be considered carefully because it affects the quality of the reconstructed signal. Evaluation from (Hohmann, 2002) suggests a system delay of 4 ms is acceptable in terms of perceptual quality. 3. GTFB in wideband audio coding PAGE 363 (a) The block diagram of the proposed perceptual audio encoder and decoder with the use of GTFB is shown in Figure 9. (b) Figure 7: (a) The envelope peak of the impulse response at channel 5 is greater than the desired delay of 4ms. (b) The delayed impulse response with its first local maximum at the desired delay position. The trade-off between an optimal system delay and the quality of the reconstructed signal needs to be considered. If the system delay is chosen to be too low, severe fluctuation of the magnitude response at lower frequencies will result. A plot of the signal to noise ratio of the reconstructed signal with respect to system delay is given in Figure 8. The SNR remains constant when the system delay is more than 12 ms, because the desired delay position is now greater than the envelope peak position of the lowest frequency channel. (a) Encoder (b) Decoder Figure 9: Block schematic of the perceptual audio coder with gammatone filterbank. (a) Encoder (b) decoder 3.1 Encoder Figure 8: The SNR of the reconstructed signal versus the system delay The synthesis technique in (Hohmann, 2002) offers a better solution than the conventional FIR/FIR or IIR/FIR analysis/synthesis filterbank as the large The input signal is a CD-quality mono digital signal, sampled at 44.1 khz with 16 bits per sample. The frame length is chosen to be 512 samples which is similar to that in MPEG-1 audio. Generally, long frame-length provides a better frequency resolution while short frame-length gives a better temporal resolution (Marina Bosi & Goldberg, 2003). The optimal frame length for the proposed coder still needs to be addressed.

The analysis GTFB consists of K channels with each gammatone filter G k (z) k=1,2, K centered at frequencies ranging from 50 Hz to 16 khz. The reason for not having gammatone filters at more than 16 khz is because the psychoacoustic model reveals that above 16 khz the masking threshold is generally greater than the signal energy. The position of the center frequency of each channel is set according to the critical bands defined in (Zwicker & Fastl, 1999). There are 24 filters used. The output from each channel is downsampled to reduce the data rate. The optimal down sampling factor for the proposed system is difficult to define since the bandwidths of the gammatone filter are non-linear and the filterbank itself is not orthogonal (Kubin & Bastiaan Kleijn, 1999). Usually in uniform filterbanks the downsampling factor is constant for every band. The aliasing term of the filter in one subband can be cancelled by another aliasing term of a filter in another subband with careful design of the synthesis filters. However for non-linear filterbanks many theoretical issues for perfect reconstruction still remain unresolved. Vaidyanathan (Akkarakaran & Vaidyanathan, 1999; Hoang & Vaidyanathan, 1989) has proposed some necessary conditions for the decimation of non-uniform filterbank but their method requires synthesis filters which is not used in our system. Therefore, the downsampling factor for our proposed system is chosen such that aliasing does not occur in the highest channel bandwidth. The highest bandwidth is 1700 Hz, so a downsampling factor of 12 is used for all channels. The psychoacoustic model from MPEG-1 Layer II is used to determine the masking threshold (ISO/IEC Int'l Standard IS 11172-3) but a 2048 point FFT is used instead of a 1024 points FFT for finer frequency resolution. The bit allocation process allocates the minimum number of bits that will be used to code each channel such that the quantization noise level is below the masking threshold. The signal level of each band is normalized between -1 and +1, and a uniform midtread quantizer is used to take into account of the zero amplitude. 6 bits are used to code the normalization factor of each channel and 4 bits are used to code the number of bits allocated to each channel, similar to that in MEPG-1 audio. 3.2 Decoder The decoder decodes and dequantizes the signal of each channel, then upsample the signal back to the input sampling frequency. Next, image extraction filters that are identical to the analysis GTFB are used to recover the frequency content of each channel. The extracted signal in each channel is then delayed by the calculated amount such that they all align at the desired delay position. A gain factor is applied to all signals before summation to make the overall system gain unity. 3.3 Subjective listening tests The result average bit rate of the system is about 1400 bits per frame, which rounds to 120 kbps. Two types of listening test were conducted. First, the coded audio signal was compared to the original CD-quality audio signal. A grade between 0 and 5 was used to measure the quality, with 0 being a disturbing degradation of the quality and 5 being unnoticeable quality change. Second, a double blind test was conducted to evaluate the quality of the proposed coder with a benchmark coder. MPEG-1 Layer III coded at 96 kbps was chosen to be the benchmark coder. Listeners were asked to select the better quality coder without the knowledge of which coder it was. 1 and 0 were given for the preferred and less-preferred coder respectively, 0.5 was given for not sure. 10 items were used in the test, which were different genres of music as well as solo instruments and vocals. 5 non-professional listeners were involved in this test. Table 1 summarises the results of the subjective listening test comparing the original and the coded audio signals. It demonstrates that the proposed coder provides adequate quality. However, some noticeable differences in quality were found with percussive instruments such as the glockenspiel and castanet. Table 2 summarises the results from the double blind test. The proposed coder demonstrates compatible quality with the MPEG-1 Layer III coded at 96 kbps for most type of music. Table 1: Quality grading of the proposed coder Test Item Average Grade Jazz 3.5 Pop 3.33 Rock 4.3 Classical 3.87 Soprano 4 Bass 2.5 Violin 3.73 Trumpet 3.3 Glockenspiel 2 Castanet 2.3 PAGE 364

Table 2: Average probability of the proposed coder preferred over MPEG-1 Layer III @ 96 kbps Test Item Average probability Jazz 0.5 Pop 0.5 Rock 0.55 Classical 0.4 Soprano 0.6 Bass 0.35 Violin 0.5 Trumpet 0.5 Glockenspiel 0.2 Castanet 0.2 4. Conclusions The bit rate of the proposed coder is higher compared to the MPEG-1 Layer III audio coder, which also achieves near transparent quality at a bit rate of 96 kbps. The high bit rate is mostly due to the limitation in decimation, as the outputs from the filterbank can not be maximally downsampled. Nonetheless, a computationally efficient synthesis technique that avoids the use of long length FIR filters is more suitable for wideband audio signals with high sampling frequencies, and a low system delay of 4 ms is attractive for application of real-time encoding/decoding 5. Acknowledgements The author would like to thank sincerely to Associate Professor E. Ambikairajah as well as Dr. Volker Hohmann for their provision of the technical papers and simulation codes. 6. References Abdulla, W. H. (2002). Auditory Based Feature Vectors for Speech Recognition Systems. Paper presented at the Advances in Communication and Software Technologies. Akkarakaran, S., & Vaidyanathan, P. P. (1999). New results and open problems on non-uniform filter-banks. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 3, 1501-1504. Ambikairajah, E., Epps, J., & Lin, L. (2001). Wideband speech and audio coding using gammatone filter banks. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2, 773-776. Ambikairajah, E., & Gunawan, T. S. (2004). Speech Enhancement using Temporal Masking and Fractional Bark Gammatone Filters. Paper presented at the International Conference on Speech Science & Technology, ASSTA Inc. Ambikairajah, E., Lin, L., & Holmes, W. H. (2004). Auditory modelling for speech processing in the perceptual domain. Australian & New Zealand Industrial and Applied Mathematics Journal (ANZIAM), 45, C960- C980. Baumgarte, F. (2002). Improved audio coding using a psychoacoustic model based on a cochlear filter bank. Speech and Audio Processing, IEEE Transactions on, 10(7), 495-503. Bosi, M. (2003). Audio Coding: Basic Principles and Recent Developments. Paper presented at the HC International Conference. Bosi, M., & Goldberg, R. E. (2003). Introduction to digital audio coding and standards. Boston: Kluwer Academic Publishers. de Boer, E. (1975). Synthetic whole-nerve action potentials for the cat. The Journal of the Acoustical Society of America, 58, 1030. Fletcher, H. (1940). Auditory Patterns. Reviews of Modern Physics, 12(1), 47-65. Hoang, P. Q., & Vaidyanathan, P. P. (1989). Non-uniform multirate filter banks: theory and design. IEEE International Symposium on Circuits and Systems, 371-374. Hohmann, V. (2002). Frequency analysis and synthesis using a Gammatone filterbank. Acta Acustica united with Acustica, 88(3), 433-442. ISO/IEC Int'l Standard IS 11172-3. Information Technology - Coding of Moving Pictures and Associated Audio for Digital Storage Media at up to about 1.5 Mbit/s - Part 3: Audio: ISO/IEC. Kubin, G., & Bastiaan Kleijn, W. (1999). On speech coding in a perceptual domain. Paper presented at the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Laakso, T. I., & Valimaki, V. (1999). Energy-Based Effective Length of the Impulse Response of a Recursive Filter. IEEE Transactions on Instrumentation and Measurement 48(1), 7. Lin, L., Ambikairajah, E., & Holmes, W. H. (2001). Auditory filter bank inversion. Paper presented at the Proc. IEEE ISCAS. Lopez-Poveda, E. A., & Meddis, R. (2001). A human nonlinear cochlear filterbank. The Journal of the Acoustical Society of America, 110, 3107. Moore, B. C. J., & Glasberg, B. R. (1983). Suggested formulae for calculating auditory-filter bandwidths and excitation patterns. The Journal of the Acoustical Society of America, 74, 750. Patterson, R. D., Robinson, K., Holdsworth, J., McKeown, D., Zhang, C., & Allerhand, M. (1992). Complex sounds and auditory images. Auditory Physiology and Perception, 429-446. Sinha, D., Johnston, J., Dorward, S., & Quackenbush, S. (1998). The perceptual audio coder (PAC) in The digital signal processing handbook. Boca Raton, Fla. New York: CRC Press, IEEE Press. Slaney, M. (1993). An efficient implementation of the Patterson-Holdsworth auditory filter bank. Perception Group, Apple Computer, Tech. Rep, 35. Zwicker, E., & Fastl, H. (1999). Psychoacoustics : facts and models (2nd updated ed.). Berlin, New York: Springer. PAGE 365