Chapter 6 Information Theory

Slides:



Advertisements
Similar presentations
Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Advertisements

Lecture 2: Basic Information Theory TSBK01 Image Coding and Data Compression Jörgen Ahlberg Div. of Sensor Technology Swedish Defence Research Agency (FOI)
Sampling and Pulse Code Modulation
Information Theory EE322 Al-Sanie.
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
Entropy Rates of a Stochastic Process
Fundamental limits in Information Theory Chapter 10 :
Chapter 11 Error-Control CodingChapter 11 : Lecture edition by K.Heikkinen.
Lecture 2: Basic Information Theory Thinh Nguyen Oregon State University.
EEE377 Lecture Notes1 EEE436 DIGITAL COMMUNICATION Coding En. Mohd Nazri Mahmud MPhil (Cambridge, UK) BEng (Essex, UK) Room 2.14.
Information Theory and Security
Noise, Information Theory, and Entropy
Noise, Information Theory, and Entropy
1 Statistical NLP: Lecture 5 Mathematical Foundations II: Information Theory.
Some basic concepts of Information Theory and Entropy
©2003/04 Alessandro Bogliolo Background Information theory Probability theory Algorithms.
Page 1 of 37 Density Evolution, Capacity Limits, and the "5k" Code Result (L. Schirber 11/22/11) The Density Evolution (DE) algorithm calculates a "threshold.
Information and Coding Theory
§1 Entropy and mutual information
STATISTIC & INFORMATION THEORY (CSNB134)
2. Mathematical Foundations
Information Theory & Coding…
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
EEET 5101 Information Theory Chapter 1
Channel Coding Part 1: Block Coding
§4 Continuous source and Gaussian channel
Gaussian Channel. Introduction The most important continuous alphabet channel is the Gaussian channel depicted in Figure. This is a time-discrete channel.
Channel Capacity
Course Review for Final ECE460 Spring, Common Fourier Transform Pairs 2.
Basic Concepts of Encoding Codes, their efficiency and redundancy 1.
Channel Capacity.
Prepared by: Amit Degada Teaching Assistant, ECED, NIT Surat
COMMUNICATION NETWORK. NOISE CHARACTERISTICS OF A CHANNEL 1.
§2 Discrete memoryless channels and their capacity function
Huffman coding Content 1 Encoding and decoding messages Fixed-length coding Variable-length coding 2 Huffman coding.
Communication System A communication system can be represented as in Figure. A message W, drawn from the index set {1, 2,..., M}, results in the signal.
Computer Vision – Compression(1) Hanyang University Jong-Il Park.
Information Theory The Work of Claude Shannon ( ) and others.
Coding Theory. 2 Communication System Channel encoder Source encoder Modulator Demodulator Channel Voice Image Data CRC encoder Interleaver Deinterleaver.
DIGITAL COMMUNICATIONS Linear Block Codes
Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
Coding Theory Efficient and Reliable Transfer of Information
CHAPTER 5 SIGNAL SPACE ANALYSIS
Additive White Gaussian Noise
Dept. of EE, NDHU 1 Chapter One Signals and Spectra.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
ECE 4710: Lecture #31 1 System Performance  Chapter 7: Performance of Communication Systems Corrupted by Noise  Important Practical Considerations: 
1 Lecture 7 System Models Attributes of a man-made system. Concerns in the design of a distributed system Communication channels Entropy and mutual information.
Jayanth Nayak, Ertem Tuncel, Member, IEEE, and Deniz Gündüz, Member, IEEE.
Jayanth Nayak, Ertem Tuncel, Member, IEEE, and Deniz Gündüz, Member, IEEE.
Source Encoder Channel Encoder Noisy channel Source Decoder Channel Decoder Figure 1.1. A communication system: source and channel coding.
Basic Concepts of Information Theory A measure of uncertainty. Entropy. 1.
1 Data Compression Hae-sun Jung CS146 Dr. Sin-Min Lee Spring 2004.
Channel Coding Theorem (The most famous in IT) Channel Capacity; Problem: finding the maximum number of distinguishable signals for n uses of a communication.
Mutual Information, Joint Entropy & Conditional Entropy
Digital Communications I: Modulation and Coding Course Spring Jeffrey N. Denenberg Lecture 3c: Signal Detection in AWGN.
Channel Coding: Part I Presentation II Irvanda Kurniadi V. ( ) Digital Communication 1.
UNIT I. Entropy and Uncertainty Entropy is the irreducible complexity below which a signal cannot be compressed. Entropy is the irreducible complexity.
UNIT –V INFORMATION THEORY EC6402 : Communication TheoryIV Semester - ECE Prepared by: S.P.SIVAGNANA SUBRAMANIAN, Assistant Professor, Dept. of ECE, Sri.
(C) 2000, The University of Michigan 1 Language and Information Handout #2 September 21, 2000.
Chapter 4: Information Theory. Learning Objectives LO 4.1 – Understand discrete and continuous messages, message sources, amount of information and its.
Basic Concepts of Information Theory Entropy for Two-dimensional Discrete Finite Probability Schemes. Conditional Entropy. Communication Network. Noise.
Increasing Information per Bit
Introduction to Information theory
Advanced Wireless Networks
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
COT 5611 Operating Systems Design Principles Spring 2012
COT 5611 Operating Systems Design Principles Spring 2014
Subject Name: Information Theory Coding Subject Code: 10EC55
Lecture 2: Basic Information Theory
Presentation transcript:

Chapter 6 Information Theory

6.1 Mathematical models for information source Discrete source

6.1 Mathematical models for information source Discrete memoryless source (DMS) Source outputs are independent random variables Discrete stationary source Source outputs are statistically dependent Stationary: joint probabilities of and are identical for all shifts m Characterized by joint PDF Model relatives to the probability, if everything is deterministic, no information is transmitted

6.2 Measure of information Entropy of random variable X A measure of uncertainty or ambiguity in X A measure of information that is required by knowledge of X, or information content of X per symbol Unit: bits (log_2) or nats (log_e) per symbol We define Entropy depends on probabilities of X, not values of X If information source is deterministic, for one value of X the probability equals 1, others equal 0, entropy of X is 0, contain no information

Shannon’s fundamental paper in 1948 “A Mathematical Theory of Communication” Can we define a quantity which will measure how much information is “produced” by a process? He wants this measure to satisfy: H should be continuous in If all are equal, H should be monotonically increasing with n If a choice can be broken down into two successive choices, the original H should be the weighted sum of the individual values of H On 3), give example on board

Shannon’s fundamental paper in 1948 “A Mathematical Theory of Communication”

Shannon’s fundamental paper in 1948 “A Mathematical Theory of Communication” The only H satisfying the three assumptions is of the form: K is a positive constant.

Binary entropy function H(p) H=0: no uncertainty H=1: most uncertainty 1 bit for binary information When equal probability, we reach the maximal – 1 bit information per symbol Probability p

Mutual information Two discrete random variables: X and Y Measures the information knowing either variables provides about the other What if X and Y are fully independent or dependent? Fully independent: I=0, fully dependent I=H(X)=H(Y)

Intuitively, if entropy H(X) is regarded as a measure of uncertainty about a random variable, then H(X|Y) is a measure of what Y does not say about X. This is "the amount of uncertainty remaining about X after Y is known",

Some properties Entropy is maximized when probabilities are equal

Joint and conditional entropy Joint entropy Conditional entropy of Y given X

Joint and conditional entropy Chain rule for entropies Therefore, If Xi are iid

6.3 Lossless coding of information source Source sequence with length n n is assumed to be large Without any source coding we need bits per symbol

Lossless source coding Typical sequence Number of occurrence of is roughly When , any will be “typical” All typical sequences have the same probability

Lossless source coding Typical sequence Since typical sequences are almost certain to occur, for the source output it is sufficient to consider only these typical sequences How many bits per symbol we need now? Number of typical sequences =

Lossless source coding Shannon’s First Theorem - Lossless Source Coding Let X denote a discrete memoryless source. There exists a lossless source code at rate R if bits per transmission This gives a fundamental limits on source coding. And entropy plays an important role in lossless compression of information source

Lossless source coding For discrete stationary source…

Lossless source coding algorithms Variable-length coding algorithm Symbols with higher probability are assigned shorter code words E.g. Huffman coding Fixed-length coding algorithm E.g. Lempel-Ziv coding

Huffman coding algorithm P(x1) P(x2) P(x3) P(x4) P(x5) x1 00 x2 01 x3 10 x4 110 x5 1110 x6 11110 x7 11111 P(x6) P(x7) Huffman coding is optimum is a sense that the average number of bits presents source symbol is a minimum, subject to that the code words satisfy the prefix condition H(X)=2.11 R=2.21 bits per symbol

6.5 Channel models and channel capacity input sequence output sequence A channel is memoryless if Channel models are useful when we design codes

Demodulator and detector Binary symmetric channel (BSC) model Source data Output data Channel encoder Binary modulator Channel Demodulator and detector Channel decoder Composite discrete-input discrete output channel

Binary symmetric channel (BSC) model 1-p p Input Output p 1 1 1-p

… … … Discrete memoryless channel (DMC) {X} {Y} Input x0 y0 Output x1 A more general case of BSC. M-ary modulation xM-1 can be arranged in a matrix yQ-1

Discrete-input continuous-output channel If N is additive white Gaussian noise… Output of detector is unquantized. Probability density function…

Discrete-time AWGN channel Power constraint For input sequence with large n

Demodulator and detector AWGN waveform channel Assume channel has bandwidth W, with frequency response C(f)=1, [-W, +W] Source data Output data Channel encoder Modulator Physical channel Demodulator and detector Channel decoder Input waveform Output waveform

AWGN waveform channel Power constraint

AWGN waveform channel How to define probabilities that characterize the channel? We can use the coefficients to characterize the channel... Equivalent to 2W uses per second of a discrete-time channel

AWGN waveform channel Power constraint becomes... Hence,

Channel capacity After source coding, we have binary sequency of length n Channel causes probability of bit error p When n->inf, the number of sequences that have np errors

Channel capacity To reduce errors, we use a subset of all possible sequences Information rate [bits per transmission] Capacity of binary channel

Channel capacity Channel encoder: add redundancy We cannot transmit more than 1 bit per channel use Channel encoder: add redundancy 2n different binary sequencies of length n contain information We use 2m different binary sequencies of length m for transmission

Channel capacity Capacity for abitray discrete memoryless channel Maximize mutual information between input and output, over all Shannon’s Second Theorem – noisy channel coding R < C, reliable communication is possible R > C, reliable communication is impossible

Channel capacity For binary symmetric channel symmetric

Channel capacity Discrete-time AWGN channel with an input power constraint For large n,

Channel capacity Discrete-time AWGN channel with an input power constraint Maximum number of symbols to transmit Transmission rate Can be obtained by directly maximizing I(X;Y), subject to power constraint

Channel capacity Band-limited waveform AWGN channel with input power constraint - Equivalent to 2W use per second of discrete-time channel bits/channel use bits/s

Channel capacity

Channel capacity Bandwidth efficiency Relation of bandwidth efficiency and power efficiency Minimum snr required for any communication system