CSI 661 - Uncertainty in A.I. Lecture 201 Basic Information Theory Review Measuring the uncertainty of an event Measuring the uncertainty in a probability.

Slides:



Advertisements
Similar presentations
DCSP-8: Minimal length coding I Jianfeng Feng Department of Computer Science Warwick Univ., UK
Advertisements

Chapter 4 Variable–Length and Huffman Codes. Unique Decodability We must always be able to determine where one code word ends and the next one begins.
EE 4780 Huffman Coding Example. Bahadir K. Gunturk2 Huffman Coding Example Suppose X is a source producing symbols; the symbols comes from the alphabet.
15-583:Algorithms in the Real World
Lecture 3: Source Coding Theory TSBK01 Image Coding and Data Compression Jörgen Ahlberg Div. of Sensor Technology Swedish Defence Research Agency (FOI)
Michael Alves, Patrick Dugan, Robert Daniels, Carlos Vicuna
Huffman code and ID3 Prof. Sin-Min Lee Department of Computer Science.
Greedy Algorithms Amihood Amir Bar-Ilan University.
Source Coding Data Compression A.J. Han Vinck. DATA COMPRESSION NO LOSS of information and exact reproduction (low compression ratio 1:4) general problem.
An introduction to Data Compression
Bounds on Code Length Theorem: Let l ∗ 1, l ∗ 2,..., l ∗ m be optimal codeword lengths for a source distribution p and a D-ary alphabet, and let L ∗ be.
SIMS-201 Compressing Information. 2  Overview Chapter 7: Compression Introduction Entropy Huffman coding Universal coding.
Data Compression.
Entropy and Shannon’s First Theorem
Lecture04 Data Compression.
Lecture 6: Huffman Code Thinh Nguyen Oregon State University.
CSc 461/561 CSc 461/561 Multimedia Systems Part B: 1. Lossless Compression.
Fundamental limits in Information Theory Chapter 10 :
DL Compression – Beeri/Feitelson1 Compression דחיסה Introduction Information theory Text compression IL compression.
Information Theory Eighteenth Meeting. A Communication Model Messages are produced by a source transmitted over a channel to the destination. encoded.
Lossless data compression Lecture 1. Data Compression Lossless data compression: Store/Transmit big files using few bytes so that the original files.
1 Chapter 5 A Measure of Information. 2 Outline 5.1 Axioms for the uncertainty measure 5.2 Two Interpretations of the uncertainty function 5.3 Properties.
CSE 326 Huffman coding Richard Anderson. Coding theory Conversion, Encryption, Compression Binary coding Variable length coding A B C D E F.
Variable-Length Codes: Huffman Codes
Fundamentals of Multimedia Chapter 7 Lossless Compression Algorithms Ze-Nian Li and Mark S. Drew 건국대학교 인터넷미디어공학부 임 창 훈.
2015/7/12VLC 2008 PART 1 Introduction on Video Coding StandardsVLC 2008 PART 1 Variable Length Coding  Information entropy  Huffman code vs. arithmetic.
Lossless Compression - I Hao Jiang Computer Science Department Sept. 13, 2007.
Information Theory and Security
Noise, Information Theory, and Entropy
Noise, Information Theory, and Entropy
Huffman Codes Message consisting of five characters: a, b, c, d,e
Basics of Compression Goals: to understand how image/audio/video signals are compressed to save storage and increase transmission efficiency to understand.
Some basic concepts of Information Theory and Entropy
Huffman Coding Vida Movahedi October Contents A simple example Definitions Huffman Coding Algorithm Image Compression.
Information and Coding Theory
Information Theory & Coding…
Richard W. Hamming Learning to Learn The Art of Doing Science and Engineering Session 10: Coding Theory I Learning to Learn The Art of Doing Science and.
Algorithms in the Real World
Source Coding-Compression
Dr.-Ing. Khaled Shawky Hassan
CS Spring 2011 CS 414 – Multimedia Systems Design Lecture 7 – Basics of Compression (Part 2) Klara Nahrstedt Spring 2011.
Lossless Compression CIS 465 Multimedia. Compression Compression: the process of coding that will effectively reduce the total number of bits needed to.
DCSP-8: Minimal length coding I Jianfeng Feng Department of Computer Science Warwick Univ., UK
Huffman coding Content 1 Encoding and decoding messages Fixed-length coding Variable-length coding 2 Huffman coding.
Huffman Code and Data Decomposition Pranav Shah CS157B.
Data Compression Meeting October 25, 2002 Arithmetic Coding.
Abdullah Aldahami ( ) April 6,  Huffman Coding is a simple algorithm that generates a set of variable sized codes with the minimum average.
CS Spring 2011 CS 414 – Multimedia Systems Design Lecture 6 – Basics of Compression (Part 1) Klara Nahrstedt Spring 2011.
Lecture 4: Lossless Compression(1) Hongli Luo Fall 2011.
Foundation of Computing Systems
Digital Image Processing Lecture 22: Image Compression
Bahareh Sarrafzadeh 6111 Fall 2009
Lossless Decomposition and Huffman Codes Sophia Soohoo CS 157B.
Chapter 7 Lossless Compression Algorithms 7.1 Introduction 7.2 Basics of Information Theory 7.3 Run-Length Coding 7.4 Variable-Length Coding (VLC) 7.5.
CS Spring 2012 CS 414 – Multimedia Systems Design Lecture 7 – Basics of Compression (Part 2) Klara Nahrstedt Spring 2012.
Huffman code and Lossless Decomposition Prof. Sin-Min Lee Department of Computer Science.
Search Engines WS 2009 / 2010 Prof. Dr. Hannah Bast Chair of Algorithms and Data Structures Department of Computer Science University of Freiburg Lecture.
ENTROPY Entropy measures the uncertainty in a random experiment. Let X be a discrete random variable with range S X = { 1,2,3,... k} and pmf p k = P X.
UNIT I. Entropy and Uncertainty Entropy is the irreducible complexity below which a signal cannot be compressed. Entropy is the irreducible complexity.
UNIT –V INFORMATION THEORY EC6402 : Communication TheoryIV Semester - ECE Prepared by: S.P.SIVAGNANA SUBRAMANIAN, Assistant Professor, Dept. of ECE, Sri.
(C) 2000, The University of Michigan 1 Language and Information Handout #2 September 21, 2000.
Information Theory Information Suppose that we have the source alphabet of q symbols s 1, s 2,.., s q, each with its probability p(s i )=p i. How much.
Ch4. Zero-Error Data Compression Yuan Luo. Content  Ch4. Zero-Error Data Compression  4.1 The Entropy Bound  4.2 Prefix Codes  Definition and.
EE465: Introduction to Digital Image Processing
Lecture 6 Instantaneous Codes and Kraft’s Theorem (Section 1.4)
Advanced Algorithms Analysis and Design
CSE 589 Applied Algorithms Spring 1999
Huffman Coding Greedy Algorithm
CSE 589 Applied Algorithms Spring 1999
Presentation transcript:

CSI Uncertainty in A.I. Lecture 201 Basic Information Theory Review Measuring the uncertainty of an event Measuring the uncertainty in a probability distribution over a set of events Other interpretations of entropy –Expected number of yes/no questions to remove uncertainty when behaving optimally/rationally. Exploring the connection that learning is compression. –So only covering lossless compression in a noise free environment.

CSI Uncertainty in A.I. Lecture 202 Coding for Compression Message : “ABBCABCB” P(A) = ¼, P(B) = ½, P(C) = ¼ How should we optimally encode the message given the above? –Use non-ambiguous set of code “words” (prefix code) –When using a binary encoding scheme Length of code to represent each symbol/event corresponds to how many yes/no questions we would need to ask to remove uncertainty of its identity. Why? –If using n-ary code, allowed to ask questions with n outcomes. Fortunately we know the formulae for H[X], set code length to be approximately equal to –log k (P(x))

CSI Uncertainty in A.I. Lecture 203 Entropies of Two Variables Joint Entropy Conditional Entropy Chain Rule Mutual Information

CSI Uncertainty in A.I. Lecture 204 Parts of the Theory of Compression To obtain efficient data compression –Problem context: What are we trying to compress –Possible data compression schemes –Measurement of how good the compression scheme is. Note, once again, average case behavior, not worst-case behavior. Formalizing the process

CSI Uncertainty in A.I. Lecture 205 Creating Decodable Codes Instantly decodable Bounded delay Uniquely decodable

CSI Uncertainty in A.I. Lecture 206 Kraft and McMillan’s Inequalities There exists an instantaneous r-ary code with code-words of length l 1 …l m iff There exists an uniquely decodable r-ary code with code-words of length l 1 …l m iff Combine the two inequalities means? This implies …

CSI Uncertainty in A.I. Lecture 207 Building an Instantaneous Prefix Code Tree Assume the Kraft inequality holds Order lengths according to l 1  …  l m Loop through symbols from shortest to longest –Allocate a node at depth l i that isn’t in a previously used subtree. –Tag all nodes “above” the selected node as unusable. Proof why this construction is always possible.

CSI Uncertainty in A.I. Lecture 208 Implication The Kraft-McMillan inequality and prefix code tree construction implies what of code-word lengths? Average code word length, average coded message length. For a given source, there maybe multiple optimal/non-optimal codes Optimal code is the minimum average code length KM Inequality implies if there is an optimal code then there is an optimal instantaneous code

CSI Uncertainty in A.I. Lecture 209 Constructing Various Codes Shannon-Fano code Huffman code –Conditions for optimality Arithmetic coding

CSI Uncertainty in A.I. Lecture 2010 Efficient Coding So given a string or body of data, we can optimally compress it. What does optimally really mean? How can we improve on optimality! When considering learning, why don’t we want to overly compress the data. How can we overcome this.