Download presentation

Presentation is loading. Please wait.

Published bySarah Rowe Modified over 4 years ago

1
DCSP-8: Minimal length coding I Jianfeng Feng Department of Computer Science Warwick Univ., UK Jianfeng.feng@warwick.ac.uk http://www.dcs.warwick.ac.uk/~feng/dsp.html

2
X=(lie on bed at 10 am today, in university at 10 am today, attend a lecture at 10 am today) p=(1/2,1/4,1/4), the entropy is ?????????/ The unit of entropy is bit/bit=bit/symbol.

3
Information source coding It seems intuitively reasonable that an information source of entropy H needs on average only H binary bits to represent each symbol. Indeed, the equi-probable BSS generate on average 1 information bit per symbol bit. However, consider the prime minister example again.

4
Suppose the probability of naked run is 0.1 (N) and that of office 0.9 (O). We have already noted that this source has an entropy of 0.47 bits/symbol. Suppose we identify naked run with 1 and office with zero.

5
This representation uses 1 binary bit per symbol, and is using more binary bits per symbol than the entropy suggests is necessary.

6
Shannon's first theorem an instantaneous code can be found that encodes a source of entropy H(X) with an average number of bits per symbol Bs such that Bs >= H(X)

7
The replacement of the symbols naked run/office with a binary representation is termed source coding. In any coding operation we replace the symbol with a codeword. The purpose of course coding is to reduce the number of bits required to convey the information provided by the information source: minimize the average length of codes.

8
Central to source coding is the use of sequence. By this, we mean that codewords are not simply associated to a single outcome, but to a sequence outcomes.

9
To see why this is useful, let us return to the problem of the prime minister. Suppose we group the outcome in three, according to theory probability, and assign binary codewords to these grouped outcomes.

10
Table 1 shows such a code, and the probability of each code word occurring. It is easy to compute that this code will on average use 1.2 bits/symbol

12
0.729log_2(0.729)+0.081log_2(0.081)*3 +0.009*log_2(0.009)*3+0.001*log_2(0.001) = -1.4070 The average length of coding is given by 0.729*1+0.081*1+2*0.081*2+2*0.009*2+3*0.009+ 3*0.001=1.2

13
This example shows how using sequences permits us to decrease the average number of bits per symbol. Moreover, without difficulty, we have found a code that has an average bit usage less than the source entropy.

14
However, there is a difficulty with the code in Table 1. Before a code word can be decoded, it must be parsed. Parsing describes that activity of breaking the message string into its component codewords.

15
After parsing, each codeword can be decoded into its symbol sequence. An instantaneously parsable code is one that can be parsed as soon as the last bit of a codeword is received.

16
An instantaneous code must satisfy the prefix condition: that no codeword may be a prefix of any other code. This condition is not satisfied by the code in Table 1.

17
Huffman coding The code in Table 1, however, is an instantaneously parsable code. It satisfies the prefix condition.

18
0.729*1+0.081*3*3+0.009*5*3+0.001*5=1. 5980

19
Decoding 1 1 1 0 1 1 0 1 0 1 1 0 0 0 0 0 0 0 0 0 1 Why do we require a code with the shortest average length?

20
The derivation of the Huffman doe tree is shown in Fig. and the tree itself is shown in Fig.. In both these figures, the letter A to H have be used in replace of the sequence in Table 2 to make them easier to read.

23
Like many theorem of information theory, the theorem tells us nothing of how to find the code. However, it is useful results.

24
For example, the code in Table 2 uses 1.6 bits/symbol which is only 0.2 bits/symbol more bits per sequence than the theorem tells us is the best we can do. We might conclude that there is little point in expending the effort in finding a code less satisfying the inequality above.

Similar presentations

OK

Problem: Huffman Coding Def: binary character code = assignment of binary strings to characters e.g. ASCII code A = 01000001 B = 01000010 C = 01000011.

Problem: Huffman Coding Def: binary character code = assignment of binary strings to characters e.g. ASCII code A = 01000001 B = 01000010 C = 01000011.

© 2018 SlidePlayer.com Inc.

All rights reserved.

To ensure the functioning of the site, we use **cookies**. We share information about your activities on the site with our partners and Google partners: social networks and companies engaged in advertising and web analytics. For more information, see the Privacy Policy and Google Privacy & Terms.
Your consent to our cookies if you continue to use this website.

Ads by Google

Ppt on layout design rules Ppt on product specification definition Ppt on different types of forests in world Ppt on bolivian water wars Ppt on quality education basketball Ppt on chapter 3 atoms and molecules for kids Ppt on different occupations pictures Ppt on internet services for class 10th Ppt on unity in diversity Ppt on weapons of mass destruction pictures