Another question consider a message (sequence of characters) from {a, b, c, d} encoded using the code shown what is the probability that a randomly chosen.

Slides:



Advertisements
Similar presentations
Theory of Discrete Information and Communication Systems Dr. Colin Campbell Room: 2.9, Queens Building 10 Lectures 2 example classes.
Advertisements

Question (from exercises 2) Are the following sources likely to be stationary and ergodic? (i)Binary source, typical sequence aaaabaabbabbbababbbabbbbbabbbbaabbbbbba......
DCSP-8: Minimal length coding II, Hamming distance, Encryption Jianfeng Feng
DCSP-10 Jianfeng Feng Department of Computer Science Warwick Univ., UK
Lecture 2: Basic Information Theory TSBK01 Image Coding and Data Compression Jörgen Ahlberg Div. of Sensor Technology Swedish Defence Research Agency (FOI)
Binary Symmetric channel (BSC) is idealised model used for noisy channel. symmetric p( 01) =p(10)
Noise, Information Theory, and Entropy (cont.) CS414 – Spring 2007 By Karrie Karahalios, Roger Cheng, Brian Bailey.
Cyclic Code.
Applied Algorithmics - week7
Error Control Code.
Sampling and Pulse Code Modulation
Probabilistic verification Mario Szegedy, Rutgers www/cs.rutgers.edu/~szegedy/07540 Lecture 4.
Information Theory EE322 Al-Sanie.
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
Chapter 6 Information Theory
Cellular Communications
Near Shannon Limit Performance of Low Density Parity Check Codes
EEE377 Lecture Notes1 EEE436 DIGITAL COMMUNICATION Coding En. Mohd Nazri Mahmud MPhil (Cambridge, UK) BEng (Essex, UK) Room 2.14.
UCB Claude Shannon – In Memoriam Jean Walrand U.C. Berkeley
Fundamental limits in Information Theory Chapter 10 :
Threshold Phenomena and Fountain Codes
DIGITAL COMMUNICATION Coding
EEE377 Lecture Notes1 EEE436 DIGITAL COMMUNICATION Coding En. Mohd Nazri Mahmud MPhil (Cambridge, UK) BEng (Essex, UK) Room 2.14.
Information Theory Eighteenth Meeting. A Communication Model Messages are produced by a source transmitted over a channel to the destination. encoded.
7/2/2015Errors1 Transmission errors are a way of life. In the digital world an error means that a bit value is flipped. An error can be isolated to a single.
Noise, Information Theory, and Entropy
CS774. Markov Random Field : Theory and Application Lecture 10 Kyomin Jung KAIST Oct
©2003/04 Alessandro Bogliolo Background Information theory Probability theory Algorithms.
INFORMATION THEORY BYK.SWARAJA ASSOCIATE PROFESSOR MREC.
DIGITAL COMMUNICATION Error - Correction A.J. Han Vinck.
1 INF244 Textbook: Lin and Costello Lectures (Tu+Th ) covering roughly Chapter 1;Chapters 9-19? Weekly exercises: For your convenience Mandatory.
Channel Coding Part 1: Block Coding
Information Coding in noisy channel error protection:-- improve tolerance of errors error detection: --- indicate occurrence of errors. Source.
Channel Capacity.
Threshold Phenomena and Fountain Codes Amin Shokrollahi EPFL Joint work with M. Luby, R. Karp, O. Etesami.
MIMO continued and Error Correction Code. 2 by 2 MIMO Now consider we have two transmitting antennas and two receiving antennas. A simple scheme called.
Redundancy The object of coding is to introduce redundancy so that even if some of the information is lost or corrupted, it will still be possible to recover.
COMMUNICATION NETWORK. NOISE CHARACTERISTICS OF A CHANNEL 1.
Data and Computer Communications by William Stallings Eighth Edition Digital Data Communications Techniques Digital Data Communications Techniques Click.
Introduction to Coding Theory. p2. Outline [1] Introduction [2] Basic assumptions [3] Correcting and detecting error patterns [4] Information rate [5]
Error Control Code. Widely used in many areas, like communications, DVD, data storage… In communications, because of noise, you can never be sure that.
Communication System A communication system can be represented as in Figure. A message W, drawn from the index set {1, 2,..., M}, results in the signal.
Outline Transmitters (Chapters 3 and 4, Source Coding and Modulation) (week 1 and 2) Receivers (Chapter 5) (week 3 and 4) Received Signal Synchronization.
DIGITAL COMMUNICATIONS Linear Block Codes
Coding Theory Efficient and Reliable Transfer of Information
1 © Unitec New Zealand CRC calculation and Hammings code.
Basic Concepts of Encoding Codes and Error Correction 1.
Error Detection and Correction – Hamming Code
1 Lecture 7 System Models Attributes of a man-made system. Concerns in the design of a distributed system Communication channels Entropy and mutual information.
Data Communications and Networking
Turbo Codes. 2 A Need for Better Codes Designing a channel code is always a tradeoff between energy efficiency and bandwidth efficiency. Lower rate Codes.
Raptor Codes Amin Shokrollahi EPFL. BEC(p 1 ) BEC(p 2 ) BEC(p 3 ) BEC(p 4 ) BEC(p 5 ) BEC(p 6 ) Communication on Multiple Unknown Channels.
INFORMATION THEORY Pui-chor Wong.
Source Encoder Channel Encoder Noisy channel Source Decoder Channel Decoder Figure 1.1. A communication system: source and channel coding.
Channel Coding Theorem (The most famous in IT) Channel Capacity; Problem: finding the maximum number of distinguishable signals for n uses of a communication.
Hamming Distance & Hamming Code
SEAC-3 J.Teuhola Information-Theoretic Foundations Founder: Claude Shannon, 1940’s Gives bounds for:  Ultimate data compression  Ultimate transmission.
Practical Session 10 Computer Architecture and Assembly Language.
Channel Coding: Part I Presentation II Irvanda Kurniadi V. ( ) Digital Communication 1.
UNIT I. Entropy and Uncertainty Entropy is the irreducible complexity below which a signal cannot be compressed. Entropy is the irreducible complexity.
UNIT –V INFORMATION THEORY EC6402 : Communication TheoryIV Semester - ECE Prepared by: S.P.SIVAGNANA SUBRAMANIAN, Assistant Professor, Dept. of ECE, Sri.
Chapter 4: Information Theory. Learning Objectives LO 4.1 – Understand discrete and continuous messages, message sources, amount of information and its.
Introduction to Information theory
Factor Graphs and the Sum-Product Algorithm
Information Theory Michael J. Watts
Chapter 6.
COT 5611 Operating Systems Design Principles Spring 2012
COT 5611 Operating Systems Design Principles Spring 2014
Information Redundancy Fault Tolerant Computing
Watermarking with Side Information
Presentation transcript:

Another question consider a message (sequence of characters) from {a, b, c, d} encoded using the code shown what is the probability that a randomly chosen bit from the encoded message is 1? a1/20 b1/410 c1/8110 d1/8111 = (expected number of 1s/expected number of bits)

Another question consider a message (sequence of characters) from {a, b, c, d} encoded using the code shown what is the probability that a randomly chosen bit from the encoded message is 1? a1/20 b1/410 c1/8110 d1/8111

Shannon-Fano theorem Channel capacity –Entropy (bits/sec) of encoder determined by entropy of source (bits/sym) –If we increase the rate at which source generates information (bits/sym) eventually we will reach the limit of the encoder (bits/sec). At this point the encoders entropy will have reached a limit This is the channel capacity S-F theorem –Source has entropy H bits/symbol –Channel has capacity C bits/sec –Possible to encode the source so that its symbols can be transmitted at up to C/H symbols per second, but no faster –(general proof in notes) source encode/ transmit receive/ decode destination channel

Conditional Entropy (lecture 3) conditional entropy of A given B=b k is the entropy of the probability distribution Pr(A|B=b k ) the conditional entropy of A given B is the average of this quantity over all b k the average uncertainty about A when B is known

Mutual information (lecture 3) H(A) – H(A|B) = H(B) – H(B|A) = H(B) + H(A|B) = H(A) +H(B|A) H(B, A ) = H(A, B) Rearrange: I(A ; B)I(B ; A) I(A ; B) = information about A contained in B H(A,B) H(A) H(B) H(A|B)H(B|A)I(A;B)

Mutual information - example A B 0 with probability p 1 with probability 1-p 0 with probability q 1 with probability 1-q C c= a+b mod 2 if p=q = 0.5, (i) what is the probability that c = 0 ? (ii) what is I(C;A) ? What if p= 0.5 and q = 0.1 ? What about the general case, any p, q ? transmit (A)receive (C) noise (B)

General case a=0 with probability p - in this case, Pr(c=0) = q, Pr(c=1) = 1-q a=1 with probability 1-p - in this case, Pr(c=0) = 1-q, Pr(c=1) = q average uncertainty about C given A = I(A;C) = H(C) - H(B) H(A|C) H(C) H(A) H(C|A) = H(B)

Discrete Channel with Noise noise channel source encode/ transmit receive/ decode dest. A X Y B A equivocation= H(X | Y ) transmission rate = H(X ) H(X | Y ) channel capacity = max (transmission rate)

Noisy Channels A noisy channel consists of an input alphabet X, an output alphabet Y and a set of conditional distributions Pr(y|x) for each y Y and x X binary symmetric channel x y

Inferring input from output x y error probability = 0.15 source distribution P(0)=0.9 observe y=1 Use Bayes x=0 is still more probable than x=1

Other useful models binary erasure channel x y Z channel x y ?

Information conveyed by a channel input distribution P(x), output distribution P(y) mutual information I(X;Y) what is the distribution P(x) that maximises I (channel capacity) I(X;Y) H(X|Y) H(Y|X) H(Y) H(X) (also depends on error matrix)

Shannons Fundamental Theorem Consider source with entropy R and a channel with capacity C such that R < C. There is a way of (block) coding the source so that it can be transmitted with arbitrarily small error group input symbols together (block code) use spare capacity for error correcting code (Hamming code etc.)

Example - noisy dice imagine restricting the input symbols to 2 and 5 This is a non-confusable subset - - for any output, we would know the input (similarly (1, 4) or {3, 6} )

Outline of proof consider sequence of signals of length N –as N increases, probability of error reduces –(typical outputs are unlikely to overlap) –as N, Pr(error) 0 –e.g. binary symmetric channel, f=0.15 repeat signal N times (figures from Mackay, Information Theory, Inference, and Learning Algorithms, CUP)

Outline consider long time T, sequence length N –2 NH(X) typical source sequences, each occurs with probability 2 -H(X) –a typical received signal y corresponds to 2 N(X|Y) possible inputs –choose 2 NR random input sequences to represent our source messages –consider transmitting x i - if it is corrupted, it may be decoded as x j where ji –if y is received, it corresponds to a set of inputs S y since R < C we can make this as small as we like by choosing large N

Error Detection / Correction noise channel source encode/ transmit receive/ decode dest. A X Y B Error code Error detect Error correct resend Error-detecting code –Detect if one or more digits has been changed –Cannot say which digits have changed –E.g. parity check Error-correcting code –Error-detection as on left –Can also work out which digits have been changed –E.g. Hamming code

Error detection If code words are very similar then it is difficult to detect errors –e.g and If code words are very different then easier to detect errors –e.g and Therefore more different code words the better –Measure using Hamming Distance, d number of different digits e.g and differ in 3 places, therefore d = Differ ?

Hamming distance Measure of distance between words Choose nearest code word e.g. a = b = c = Use d to predict number of errors we can detect/correct e.g. parity check sent rec rec. d 2e+1 Can correct up to e errors per word d = 2eCan correct up to e-1 errors per word, can detect e errors d e+1 Can detect up to e errors per word

Error correction Like error-detecting code, but need more bits (obvious really!). More efficient when larger code words are being used Overhead of coding/decoding arithmetic Hamming code –D = number of data bits –P = number of parity bits –C = D + P = code word length –Hamming inequality for single error correction: D + P P –If P is small – hardly worth doing – cheaper to re-send code word –If P 3 some increase in transmission rate is possible

Hamming code Process: –Coding Take code word from encoder (before adding parity bits) and multiply by generator matrix G using modulo-2 arithmetic This gives the code word –(d 1, …, d D, p 1, …, p P ) –Decoding Take received code word (D+P bits) and multiply by decoder matrix X using modulo-2 arithmetic This gives us a syndrome (or parity) vector s If s contains all zeros then no errors Vector s is matched against X to find position of single error.

Example E.g. d = 4, p = 3 G = [ I | A ] Encode 1001 X = [A T | I] Receive & decode Error in bit 2

Summary probability joint conditional bayes entropy decomposition conditional mututal information sources simple Markov stationary ergodic information capacity source coding theorem coding optimal / compression channel capacity shannons fundamental theorem error correction/detection noise channel source encode/ transmit receive/ decode dest. A X Y B

Next stop … Theory of Discrete Information and Communication Systems weeks 1-6 Communications Systems Performance (Mark Beach) Coursework Languages, Automata and Complexity (Colin Campbell) 1st order Predicate Logic (Enza di Tomaso) EMAT Information Systems (CSE 3, Eng Maths 3, Knowledge Eng 3) EMAT Logic and Information (CSE 2, Eng Maths 2) EENG Communication Systems (EE 3, Avionics 3) EENG M2100 Communication Systems (MSc Comms/Sig Proc) weeks 7-12

Pictorial representation I(X;Y) H(X|Y) H(Y|X) H(Y) H(X) (from Volker Kuhn, Bremen)

error correcting - transmitted bits - choose st even parity in each set s - source bits any two 4-bit codewords differ in at least 3 places

xi xj

graph