1 / 30

Introduction of MPEG-2 AAC Audio Coding

Introduction of MPEG-2 AAC Audio Coding. 指導教授 : 蔡宗漢 學生 : 劉俊男. Why do we need MPEG?. low sample and bit rates storage space For example: A CD can contain a maximum of 650 MB of unencoded video just 5 or 6 minutes.

binh
Download Presentation

Introduction of MPEG-2 AAC Audio Coding

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Introduction of MPEG-2 AAC Audio Coding 指導教授:蔡宗漢 學生:劉俊男

  2. Why do we need MPEG? • low sample and bit rates • storage space • For example: • A CD can contain a maximum of 650 MB of unencoded video just 5 or 6 minutes. • When the video signal is encoded the CD can contain up to 74 minutes of video. • Bandwidths

  3. MPEG Audio Coding Standards • MPEG-1 (1992) • Three layers with increasing complexity and performance • Layer-3 is the highest complexity mode,optimized to provide the highest quality at low bitrate (around 128 kbits/s for a stereo signal) • MPEG-2 (1994) • backwards compatible multichannel coding • coding at lower sampling frequencies adds sampling frequencies of 16,22.05,24 khz • MPEG-2 AAC (1994) • AAC is a second generation audio coding scheme for generic coding of stereo and multichannel signals

  4. MPEG Audio Coding Standards • MPEG-4 (1998) • the emphasis in MPEG-4 is on new functionalities rather than better compression efficiency • mobile as well as stationary user terminals,database access, communications,will be major applications for MPEG-4 • consists of a family of audio coding algorithm spanning the range from low bitrate speech coding (down to 2 kbit/s) up to high quality audio coding at 64 kbit/s per channel and above. • generic audio coding at medium to high bitrate is down by AAC • MPEG-7 (2001) • does not define compression algorithms • MPEG-7 is a content representation standard for multimedia information search,filtering,management and processing

  5. low medium high Channel bitrate(kbps) 2 4 6 8 10 12 14 16 24 32 48 64 ~ Scalable Coder Parametric coder CELP coder ~ T/F coder ITU-T coder Signal Bandwidth 4 kHz 8 kHz 20 kHz Assignment of codecs to bitrate ranges in MPEG-4 natural audio coding

  6. Structure of MPEG-4 Audio

  7. Audio in Analysis Filterbank Quantization & Coding Encoding of bitstream bistream out Perceptual Model Audio out bistream in Decoding of bitstream Inverse Quantization Synthesis Filterbank A basic perceptual audio coder Block diagram of a perceptual Encoding system Block diagram of a perceptual Decoding system

  8. The critical bands

  9. The absolute threshold of hearing in quiet • Across the audio spectrum, quantifies sound pressure level (SPL) required at each frequency such that an average listener will detect a pure tone stimulus in a noiseless environment

  10. The absolute threshold of hearing in quiet • The absolute threshold of hearing characterizes the amount of energy needed in a pure tone such that it can be detected by a listener in a noiseless environment. • The absolute threshold is typically expressed in terms of dB Sound Pressure Level (dB SPL). • The quiet threshold is well approximated by the non-linear function

  11. SIMULTANEOUS MASKING

  12. TEMPORAL MASKING • Pre-masking in particular has been exploited in conjunction with adaptive block size transform coding to compensate for pre-echo distortions.

  13. Pre-echo effect • Pre-Echo Example: (a) Uncoded Castanets. (b) Transform Coded Castanets, 2048-Point Block Size

  14. An iterative method is • employed • So as to keep the quantization • noise in all critical bands • below the global masking • threshold The building blocks of MPEG-2 AAC encoder • A high frequency resolution • filterbank (MDCT) • Switched between resolutions • of 1024 and 128 spectral lines • The shape of the transform • window can be adaptively • selected between a sine • window an a Kaiser-Bessel- • derived(KBD) window • Depending on the stationary • or transient character of the • input signal The temporal noise shaping tool controls the time dependence of the quantization noise the perceptual model is taken from MPEG-1(model 2). • The second-order backward • adaptive predictor • Improves coding efficiency .

  15. MPEG-2 AAC Decoder

  16. Channel mapping • supports up to 46 channels for various multichannel loudspeaker configurations and other applications • the default loudspeaker configurations are • the monophonic channel • the stereophonic channel • the 5.1 system (five channels plus LFE channel).

  17. Applications for MPEG-2 AAC • Due to its high coding efficiency, AAC is a prime candidate for any digital broadcasting system. • The Japanese authorities were the first to decide to use AAC within practically all digital audio broadcasting schemes. As their first services will start in the year 2000, this decision already triggered the development of dedicated AAC decoder chips at a number of manufacturers. • AAC has been selected for the use within the Digital Radio Mondiale (DRM) system. Due to its superior performance, AAC will also play a major role for the delivery of high-quality musicvia the Internet.

  18. Applications for MPEG-2 AAC • Furthermore, AAC (with some modifications) is the only high-quality audio coding scheme used within the MPEG-4 standard, the future "global multimedia language". • Fraunhofer IIS-A offers to contribute to AAC applications at all implementation levels, e.g. licensing software libraries for PC-based applications or for VLSI developments as well as offering DSP-based solutions (e.g. on Motorola’s DSP56300, Texas Instruments’ TMS320C67xx, and Analog Devices’ ADSP21x6x family). The coding methods developed by Fraunhofer IIS-A stand for optimum audio quality at any given bit rate.

  19. Profiles of MPEG-2 AAC (1) main profile •offers highest quality • used when memory cost is not significant • substantial processing power is available (2) low-complexity profile (LC) • used when RAM usage, processing power and compression requirements are all present • preprocessing and time-domain prediction are not permitted • TNS order and bandwidth are limited (3) scaleablesampling rate profile (SSR) • offers the lowest complexity • preprocessing block is added,and prediction is not permitted •TNS order and bandwidth are limited

  20. Tool usage of AAC Profiles • Profile Interoperability

  21. MPEG-2 AAC audio transport formats • the basic audio format and the transport syntax for synchronization and coding parameters in MPEG-1 are tied together unseparably • MPEG-2 AAC defines both,but leaves the actual choice of audio transport syntax to the application • ADIF (Audio Data Interchange Format) • puts all data controlling the decoder (like sampling frequency, mode etc.) into a single header preceding the actual audio stream • it is useful for file exchange , but does not allow for break-in or start of decoding at any point in time like the MPEG-1 format • ADTS (Audio Data Transport Stream) • format packs AAC data into frames with headers very similar to the MPEG-1 header format • allows start of decoding in the middle of an audio bitstream • the ADTS format has emerged as the de-facto standard for a number of applications using AAC

  22. Filterbank and block switching • Standard Filterbank • A straight forward Modified Discrete Cosine Transform (MDCT) • Supporting block lengths of 2048 points and 256 points which can be switched dynamically • Supports two different window shapes that can be switched dynamically • sine shaped window • Kaiser-Bessel Derived (KBD) Window • All blocks are overlapped by 50% with the preceding and the following block

  23. Frame k Frame k+1 Frame k+2 Frame k+3 M M M M MDCT MDCT MDCT MDCT M M MDCT M MDCT M M M 2M 2M 2M 2M 2M 2M + + ‧ ‧ ‧ ‧ ‧ ‧ M M Frame k+1 Frame k+2 MDCT & IMDCT • the MDCT basis functions extend across two blocks in time, leading to virtual elimination of the blocking artifacts

  24. Block switching and Overlap-add 

  25. Temporal noise shaping (TNS)  • The basic idea of TNS relies on the duality of time and frequency domain • TNS uses a prediction approach in the frequency domain to shape the quantization noise over time • It applies a filter to the original spectrum and quantizes this filtered signal • quantized filter coefficients are transmitted in the bitstream • the decoder undo the filtering performed in the encoder, leading to a temporally shaped distribution of quantization noise in the decoded audio signal

  26. Frequency domain prediction • Improves redundancy reduction of stationary signal segments • Only supported in AAC Main • The actual implementation of the predictor is a second order backwards adaptive lattice structure • The required processing power of the frequency domain prediction and the sensitivity to numerical imperfections make this tool hard to use on fixed point platforms

  27. Joint stereo coding • Mid-Side (MS) stereo coding • Applies a matrix to the left and right channel signals, computing sum and difference of the two original signals • Intensity stereo coding • Saving bitrate by replacing the left and the right signal by a single representing signal plus directional information • Intensity stereo is by definition a lossy coding method thus it is primarily useful at low bitrates. For coding at higher bitrates only MS stereo is used.

  28. Scalefactors • Inherent noise shaping in the non-linear quantizer is usually not sufficient to achieve acceptable audio quality • Scalefactors are used to amplify the signal in certain spectral regions (the scalefactor bands) to increase the signal-to-noise ratio in these bands • To properly reconstruct the original spectral values in the decoder the scalefactors have to be transmitted within the bitstream • Scalefactors are coded as efficiently as possible • differentially encoded and then Huffman

  29. Quantization • A non-linear quantizer is used • The main source of the bitrate reduction • It assignes a bit allocation to the spectral values according to the accuracy demands determined by the perceptual model • The main advantage over a conventional linear quantizer is the implicit noise shaping

  30. 11 huffman codebooks for the spectral data 2 huffman codebooks for the intensity stereo Noiseless coding • The noiseless coding tries to optimize the redundancy reduction within the spectral data coding • The spectral data is encoded using a Huffman code Neither spectral coefficients nor a scalefactor transmitted

More Related