Download presentation

Presentation is loading. Please wait.

Published byAnnika Barwick Modified over 2 years ago

1
Information & Entropy

2
Shannon Information Axioms Small probability events should have more information than large probabilities. – “the nice person” (common words lower info) – “philanthropist” (less used more information) Information from two disjoint events should add – “engineer” Information I 1 – “stuttering” Information I 2 – “stuttering engineer” Information I 1 + I 2

3
Shannon Information p I

4
Information Units log 2 – bits log e – naps log 10 – ban or a hartley Ralph Vinton Lyon Hartley (1888-1970) inventor of the electronic oscillator circuit that bears his name, a pioneer in the field of Information Theory

5
Illustration Q: We flip a coin 10 times. What is the probability we come up the sequence 0 0 1 1 0 1 1 1 0 1? Answer How much information do we have?

6
Illustration: 20 Questions Interval halving: Need 4 bits of information

7
Entropy Bernoulli trial with parameter p Information from a success = Information from a failure = (Weighted) Average Information Average Information = Entropy

8
The Binary Entropy Function p

9
Entropy Definition =average Information

10
Entropy of a Uniform Distribution

11
Entropy as an Expected Value where

12
Entropy of a Geometric RV then H = 2 bits when p =0.5

13
Relative Entropy

14
Relative Entropy Property Equality iff p=q

15
Relative Entropy Property Proof Since

16
Uniform Probability is Maximum Entropy Relative to uniform: Thus, for K fixed, How does this relate to thermodynamic entropy?

17
Entropy as an Information Measure: Like 20 Questions 16 Balls Bill Chooses One 1111 22 33 22 44 7658 You must find which ball with binary questions. Minimize the expected number of questions.

18
One Method... 1234658 yes no yes no yes no yes no yes no yes no 7

19
Another (Better) Method... yes no 123465 87 yes no yes no Longer paths have smaller probabilities. 1111 22 33 22 44 7658

20
yes no 123465 87 yes no yes no 1111 22 33 22 44 7658

21
Relation to Entropy... The Problem’s Entropy is... 1111 22 33 22 44 7658

22
Principle... The expected number of questions will equal or exceed the entropy. There can be equality only if all probabilities are powers of ½. 1111 22 33 22 44 7658 1111 22 33 22 44 7658 1111 22 33 22 44 7658 1111 22 33 22 44 7658

23
Principle Proof 1111 22 33 22 44 7658 Lemma: If there are k solutions and the length of the path to the k th solution is, then

24
Principle Proof = the relative entropy with respect to Since the relative entropy always is nonnegative...

Similar presentations

OK

Expected values of discrete Random Variables. The function that maps S into S X in R and which is denoted by X(.) is called a random variable. The name.

Expected values of discrete Random Variables. The function that maps S into S X in R and which is denoted by X(.) is called a random variable. The name.

© 2018 SlidePlayer.com Inc.

All rights reserved.

Ads by Google

Ppt on microcontroller based digital thermometer Ppt on next generation 2-stroke engine problems Convert word doc to ppt online training By appt only movie Ppt on solar based traffic light control system Ppt on excess demand and deficient demand Maths ppt on real numbers class 10 Ppt on air pollution act Download ppt on nutrition in plants and animals Ppt on rulers and buildings