Compression No. 1  Seattle Pacific University Data Compression Kevin Bolding Electrical Engineering Seattle Pacific University.

Slides:



Advertisements
Similar presentations
T.Sharon-A.Frank 1 Multimedia Compression Basics.
Advertisements

15 Data Compression Foundations of Computer Science ã Cengage Learning.
Data Compression CS 147 Minh Nguyen.
An introduction to Data Compression
Speech Compression. Introduction Use of multimedia in personal computers Requirement of more disk space Also telephone system requires compression Topics.
Digital Representation of Audio Information Kevin D. Donohue Electrical Engineering University of Kentucky.
1 Digital Audio Compression. 2 Formats  There are many different formats for storing and communicating digital audio:  CD audio  Wav  Aiff  Au 
Speech Group INRIA Lorraine
CELLULAR COMMUNICATIONS 5. Speech Coding. Low Bit-rate Voice Coding  Voice is an analogue signal  Needed to be transformed in a digital form (bits)
Speech Coding Nicola Orio Dipartimento di Ingegneria dell’Informazione IV Scuola estiva AISV, 8-12 settembre 2008.
1 CODING AND COMPRESSION PRESENTED BY: PING CHEN CECS401 UMC DATE: April,
1 Audio Compression Techniques MUMT 611, January 2005 Assignment 2 Paul Kolesnik.
Spatial and Temporal Data Mining
EE2F1 Speech & Audio Technology Sept. 26, 2002 SLIDE 1 THE UNIVERSITY OF BIRMINGHAM ELECTRONIC, ELECTRICAL & COMPUTER ENGINEERING Digital Systems & Vision.
COMP 249 :: Spring 2005 Slide: 1 Audio Coding Ketan Mayer-Patel.
EE2F1 Speech & Audio Technology Sept. 26, 2002 SLIDE 1 THE UNIVERSITY OF BIRMINGHAM ELECTRONIC, ELECTRICAL & COMPUTER ENGINEERING Digital Systems & Vision.
1 Lossless Compression Multimedia Systems (Module 2) r Lesson 1: m Minimum Redundancy Coding based on Information Theory: Shannon-Fano Coding Huffman Coding.
1 Audio Compression Multimedia Systems (Module 4 Lesson 4) Summary: r Simple Audio Compression: m Lossy: Prediction based r Psychoacoustic Model r MPEG.
Image Compression - JPEG. Video Compression MPEG –Audio compression Lossy / perceptually lossless / lossless 3 layers Models based on speech generation.
CS :: Fall 2003 Audio Coding Ketan Mayer-Patel.
Speech coding. What’s the need for speech coding ? Necessary in order to represent human speech in a digital form Applications: mobile/telephone communication,
Media File Formats Jon Ivins, DMU. Text Files n Two types n 1. Plain text (unformatted) u ASCII Character set is most common u 7 bits are used u This.
A/D Conversion No. 1  Seattle Pacific University Analog to Digital Conversion Based on Chapter 5 of William Stallings, Data and Computer Communication.
Digital Audio Watermarking: Properties, characteristics of audio signals, and measuring the performance of a watermarking system نيما خادمي کلانتري
Lecture 10 Data Compression.
Speech Coding PCM DPCM ADPCM LPC CELP A road map Page 1 of 30
1 ELEN 6820 Speech and Audio Processing Prof. D. Ellis Columbia University Midterm Presentation High Quality Music Metacompression Using Repeated- Segment.
Speech Coding Using LPC. What is Speech Coding  Speech coding is the procedure of transforming speech signal into more compact form for Transmission.
Acoustic Analysis of Speech Robert A. Prosek, Ph.D. CSD 301 Robert A. Prosek, Ph.D. CSD 301.
ECE 4710: Lecture #9 1 PCM Noise  Decoded PCM signal at Rx output is analog signal corrupted by “noise”  Many sources of noise:  Quantizing noise »Four.
10/10/04 L5/1/28 COM342 Networks and Data Communications Ian McCrumRoom 5D03B Tel: voice.
Page 0 of 23 MELP Vocoders Nima Moghadam SN#: Saeed Nari SN#: Supervisor Dr. Saameti April 2005 Sharif University of Technology.
Speech Coding Submitted To: Dr. Mohab Mangoud Submitted By: Nidal Ismail.
CE Digital Signal Processing Fall 1992 Waveform Coding Hossein Sameti Department of Computer Engineering Sharif University of Technology.
Compression.  Compression ratio: how much is the size reduced?  Symmetric/asymmetric: time difference to compress, decompress?  Lossless; lossy: any.
1 PCM & DPCM & DM. 2 Pulse-Code Modulation (PCM) : In PCM each sample of the signal is quantized to one of the amplitude levels, where B is the number.
1 Audio Compression. 2 Digital Audio  Human auditory system is much more sensitive to quality degradation then is the human visual system  redundancy.
Lossless Compression CIS 465 Multimedia. Compression Compression: the process of coding that will effectively reduce the total number of bits needed to.
Submitted By: Santosh Kumar Yadav (111432) M.E. Modular(2011) Under the Supervision of: Mrs. Shano Solanki Assistant Professor, C.S.E NITTTR, Chandigarh.
ECE 5525 Osama Saraireh Fall 2005 Dr. Veton Kepuska
VOCODERS. Vocoders Speech Coding Systems Implemented in the transmitter for analysis of the voice signal Complex than waveform coders High economy in.
Advances in digital image compression techniques Guojun Lu, Computer Communications, Vol. 16, No. 4, Apr, 1993, pp
Digital Multiplexing 1- Pulse Code Modulation 2- Plesiochronous Digital Hierarchy 3- Synchronous Digital Hierarchy.
ITU-T G.729 EE8873 Rungsun Munkong March 22, 2004.
Additive White Gaussian Noise
Dr. Sudharman K. Jayaweera and Amila Kariyapperuma ECE Department University of New Mexico Ankur Sharma Department of ECE Indian Institute of Technology,
Digital Audio III. Sound compression (I) Compression of sound data requires different techniques from those for graphical data Requirements are less stringent.
1 Audio Coding. 2 Digitization Processing Signal encoder Signal decoder samplingquantization storage Analog signal Digital data.
CS654: Digital Image Analysis Lecture 34: Different Coding Techniques.
Performance Comparison of Speaker and Emotion Recognition
Chapter 1 Background 1. In this lecture, you will find answers to these questions Computers store and transmit information using digital data. What exactly.
Voice Sampling. Sampling Rate Nyquist’s theorem states that a signal can be reconstructed if it is sampled at twice the maximum frequency of the signal.
ECE 101 An Introduction to Information Technology Information Coding.
1 Speech Compression (after first coding) By Allam Mousa Department of Telecommunication Engineering An Najah University SP_3_Compression.
Submitted To-: Submitted By-: Mrs.Sushma Rani (HOD) Aashish Kr. Goyal (IT-7th) Deepak Soni (IT-8 th )
Data Compression.
Digital Communications Chapter 13. Source Coding
Vocoders.
Data Compression.
Data Compression CS 147 Minh Nguyen.
1 Vocoders. 2 The Channel Vocoder (analyzer) : The channel vocoder employs a bank of bandpass filters,  Each having a bandwidth between 100 HZ and 300.
Why Compress? To reduce the volume of data to be transmitted (text, fax, images) To reduce the bandwidth required for transmission and to reduce storage.
Linear Predictive Coding Methods
Mobile Systems Workshop 1 Narrow band speech coding for mobile phones
UNIT IV.
Image Transforms for Robust Coding
Speech coding.
15 Data Compression Foundations of Computer Science ã Cengage Learning.
Govt. Polytechnic Dhangar(Fatehabad)
15 Data Compression Foundations of Computer Science ã Cengage Learning.
Presentation transcript:

Compression No. 1  Seattle Pacific University Data Compression Kevin Bolding Electrical Engineering Seattle Pacific University

Compression No. 2  Seattle Pacific University Compression Reducing the size of a message by removing redundant data Compression relies on some understanding of the structure of a message Only works on structured messages; random messages cannot be effectively compressed Because wireless systems (especially mobile phones) have limited capacity, compression is critical For example, an average compression rate of 50% doubles system capacity

Compression No. 3  Seattle Pacific University Lossless compression Lossless compression transmits enough information to recreate the original with no defects Huffman coding Calculate statistics for the occurrence rates of various symbols Assign short codes to common signals, long codes to uncommon symbols

Compression No. 4  Seattle Pacific University Huffman Encoding Example: Alphabet of {A,B,C,D,E,F,G} We could pick A=000, B=001, C=010, etc. Encoding of the string: BACADAEAFABBAAAGAH is 54 bits: What if we know that some symbols occur more frequently (A, followed by B in this example)? A=0; B=100; C=1010; D=1011; E=1100; F=1101; G=1110; H=1111 Encoding of the previous string is only 42 bits:

Compression No. 5  Seattle Pacific University General Lossless Compression If source type is known (i.e. English text), a pre- defined dictionary/index can be used If source type is not known, the algorithm can create a dictionary with statistics on the fly Dictionary must be transmitted with the message ZIP file compression works this way

Compression No. 6  Seattle Pacific University Lossy Compression If the only purpose of the message is to be interpreted by a human, we can remove data not perceived by human senses Music – Limits on hearing Cannot hear frequencies above 20kHz Cannot detect frequencies with low amplitudes in the presence of frequencies with high amplitudes Voice – Purpose is communication, not fidelity Intelligibility matters most Pleasing sound also matters Images and Video Models of how humans see can help…

Compression No. 7  Seattle Pacific University Voice Compression Tremendous opportunity for compression We know the model for voice production We know the model for reception We probably know the rules of speech (language- specific) Language has tremendous redundancy Speech is time-limited, with communication as its primary purpose; it is not generally saved and studied in more detail later

Compression No. 8  Seattle Pacific University Voice Compression Techniques Companding Use non-linear encoding to decrease dynamic range Reduces the number of bits needed per sample Linear Predictive Coding Use a model of the human vocal tract Send parameters that reflect operations on the components of the vocal tract Codebooks Form a book of “codes” that correspond to types of sounds commonly used in speed Send the number for each code rather than the sound itself

Compression No. 9  Seattle Pacific University Nonlinear coding Linear coding samples using a linear scale Large amplitude  Cross many levels during waveform (good) Nonlinear encoding More levels concentrated near the center level Wider spacing at edges Compresses dynamic range Small amplitude  Cross few levels during waveform (bad)

Compression No. 10  Seattle Pacific University Companding A form of non-linear encoding commonly used with telephone systems Before quantitizing, reshape the signal by increasing the amplitude of low-amplitude signals In US and Japan, mu-law is used; in Europe, A-law is used Original (top) and companded (bottom) waveforms.

Compression No. 11  Seattle Pacific University Linear Predictive Coding The human vocal tract can be modeled by: A buzzer with varying pitch and volume (the vocal cords) A tube with resonances at certain frequencies (the vocal tract) Resonant frequencies are called formants A filter (lips, tongue, etc.) LPC Encoding Estimate the frequency and intensity of the original buzz Find the closest match to overall tone by selecting formants Remove these from signal  left-over is residue Transmit the buzz frequency and intensity, the formants selected, and the residue

Compression No. 12  Seattle Pacific University Codebooks Since speech is chosen from a limited number of basic sounds, it may be described as only these basic sounds Make a codebook including examples of all types of speech sounds Encoding: Find best matching sound and send the code for that Decoding: Look up code and produce sound Drawbacks: Codebook may reflect “standard” speech, not the speaker in particular Codebook may be excessively large

Compression No. 13  Seattle Pacific University Codebook Excited Linear Predictive Coders CELP coders use two steps Encode the speech using an LPC. This gives codes for the frequency, intensity and formants, plus a residue. Encode the residue using codebook techniques. Normally, two codebooks are used: Fixed codebook of common residue signals. Adaptive codebook that includes recently used residue signals. This reduces the data rate for repetitive signals.