TSBK01 Image Coding and Data Compression Lecture 2: Basic Information Theory Jörgen Ahlberg Div.
Download ReportTranscript TSBK01 Image Coding and Data Compression Lecture 2: Basic Information Theory Jörgen Ahlberg Div.
TSBK01 Image Coding and Data Compression Lecture 2: Basic Information Theory Jörgen Ahlberg Div. of Sensor Technology Swedish Defence Research Agency (FOI) Today 1. 2. 3. 4. 5. What is information theory about? Stochastic (information) sources. Information and entropy. Entropy for stochastic sources. The source coding theorem. Part 1: Information Theory The Claude Shannon: A Mathematical Theory of Communication Bell System Technical Journal, 1948 Sometimes referred to as ”Shannon-Weaver”, since the standalone publication has a foreword by Weaver. Be careful! Quotes about Shannon ”What is information? Sidestepping questions about meaning, Shannon showed that it is a measurable commodity”. ”Today, Shannon’s insight help shape virtually all systems that store, process, or transmit information in digital form, from compact discs to computers, from facsimile machines to deep space probes”. ”Information theory has also infilitrated fields outside communications, including linguistics, psychology, economics, biology, even the arts”. Change to an efficient Change representation, to an efficient representation for, Any source of information i.e., data compression. transmission, i.e., error control coding. Source Channel Source Channel coder coder Channel Channel Source Sink, decoder decoder receiver The channel is anything transmitting or storing information – Recover from channel distortion.Uncompress a radio link, a cable, a disk, a CD, a piece of paper, … Fundamental Entities Source C H Channel Source coder R Channel coder Channel C Channel Source Sink, decoder decoder receiver H: The information content of the source. R: Rate from the source coder. C: Channel capacity. Fundamental Theorems Source C H Channel Source coder R Channel coder Channel C Channel Source Sink, decoder decoder receiver Shannon 1: Error-free transmission possible if R¸H and C¸R. Shannon 2: Source coding and channel coding can be Source coding Channel theorem coding (simplified) theorem (simplified) optimized independently, and binary symbols can be used as intermediate format. Assumption: Arbitrarily long delays. Part 2: Stochastic sources A source outputs symbols X1, X2, ... Each symbol take its value from an alphabet Example 1: A text is a sequence of symbols A (digitized) grayscale image is a A Example = (a1, a2,2:…). each taking its value from the alphabet sequence of symbols each taking its value Model: P(X ) assumed be known for A = (a, …, z,1,…,X A, …,NZ, 1, 2, …9, !,to ?, …). from the alphabet A = (0,1) or A = (0, …, 255). all combinations. Source X1, X2, … Two Special Cases 1. The Memoryless Source Each symbol independent of the previous ones. P(X1, X2, …, Xn) = P(X1) ¢ P(X2) ¢ … ¢ P(Xn) 2. The Markov Source Each symbol depends on the previous one. P(X1, X2, …, Xn) = P(X1) ¢ P(X2|X1) ¢ P(X3|X2) ¢ … ¢ P(Xn|Xn-1) The Markov Source A symbol depends only on the previous symbol, so the source can be modelled by a state diagram. 0.7 b 0.5 a 1.0 0.3 0.2 c 0.3 A ternary source with alphabet A = (a, b, c). The Markov Source Assume we are in state a, i.e., Xk = a. The probabilities for the next symbol are: 0.7 b P(Xk+1 = a | Xk = a) = 0.3 0.5 a 1.0 0.3 0.2 P(Xk+1 = b | Xk = a) = 0.7 P(Xk+1 = c | Xk = a) = 0 c 0.3 The Markov Source So, if Xk+1 = b, we know that Xk+2 will equal c. 0.7 b P(Xk+2 = a | Xk+1 = b) = 0 0.5 a 1.0 0.3 0.2 P(Xk+2 = b | Xk+1 = b) = 0 P(Xk+2 = c | Xk+1 = b) = 1 c 0.3 The Markov Source If all the states can be reached, the stationary probabilities for the states can be calculated from the given transition probabilities. Stationary probabilities? That’s the Markov models can be usedto represent probabilities i = P(Xk = ai) for any k when Xk-1, Xmore not one given. sources with dependencies than k-2, … are step back. – Use a state diagram with several symbols in each state. Analysis and Synthesis Stochastic models can be used for analysing a source. – Find a model that well represents the real-world source, and then analyse the model instead of the real world. Stochastic models can be used for synthesizing a source. – Use a random number generator in each step of a Markov model to generate a sequence simulating the source. Show plastic slides! Part 3: Information and Entropy Assume a binary memoryless source, e.g., a flip of a coin. How much information do we receive when we are told that the outcome is heads? – If it’s a fair coin, i.e., P(heads) = P (tails) = 0.5, we say that the amount of information is 1 bit. – If we already know that it will be (or was) heads, i.e., P(heads) = 1, the amount of information is zero! – If the coin is not fair, e.g., P(heads) = 0.9, the amount of information is more than zero but less than one bit! – Intuitively, the amount of information received is the same if P(heads) = 0.9 or P (heads) = 0.1. Self Information So, let’s look at it the way Shannon did. Assume a memoryless source with – alphabet A = (a1, …, an) – symbol probabilities (p1, …, pn). How much information do we get when finding out that the next symbol is ai? According to Shannon the self information of ai is Why? Assume two independent events A and B, with probabilities P(A) = pA and P(B) = pB. For both the events to happen, the probability is pA ¢ pB. However, the amount of information should be added, not multiplied. Logarithms satisfy this! No, we want the information to increase with decreasing probabilities, so let’s use the negative logarithm. Self Information Example 1: Example 2: Which logarithm? Pick the one you like! If you pick the natural log, you’ll measure in nats, if you pick the 10-log, you’ll get Hartleys, if you pick the 2-log (like everyone else), you’ll get bits. Self Information On average over all the symbols, we get: H(X) is called the first order entropy of the source. This can be regarded as the degree of uncertainty about the following symbol. Entropy Example: Binary Memoryless Source 01101000… BMS Let Then 1 The uncertainty (information) is greatest when 0 0.5 1 Entropy: Three properties 1. It can be shown that 0 · H · log N. 2. Maximum entropy (H = log N) is reached when all symbols are equiprobable, i.e., pi = 1/N. 3. The difference log N – H is called the redundancy of the source. Part 4: Entropy for Memory Sources Assume a block of source symbols (X1, …, Xn) and define the block entropy: That is, the summation is done over all possible combinations of n symbols. The entropy for a memory source is defined as: That is, let the block length go towards infintity. Divide by n to get the number of bits / symbol. Entropy for a Markov Source The entropy for a state Sk can be expressed as Pkl is the transition probability from state k to state l. Averaging over all states, we get the entropy for the Markov source as The Run-length Source Certain sources generate long runs or bursts of equal symbols. Example: 1- A B 1- Probability for a burst of length r: P(r) = (1-)r-1¢ Entropy: HR = - r=11 P(r) log P(r) If the average run length is , then HR/ = HM. Part 5: The Source Coding Theorem The entropy is the smallest number of bits allowing error-free representation of the source. Why is this? Let’s take a look on typical sequences! Typical Sequences Assume a long sequence from a binary memoryless source with P(1) = p. Among n bits, there will be approximately w = n ¢ p ones. Thus, there is M = (n over w) such typical sequences! Only these sequences are interesting. All other sequences will appear with smaller probability the larger is n. How many are the typical sequences? bits/symbol Enumeration needs log M bits, i.e, bits per symbol! How many bits do we need? Thus, we need H(X) bits per symbol to code any typical sequence! The Source Coding Theorem Does tell us – that we can represent the output from a source X using H(X) bits/symbol. – that we cannot do better. Does not tell us – how to do it. Summary The mathematical model of communication. – Source, source coder, channel coder, channel,… – Rate, entropy, channel capacity. Information theoretical entities – Information, self-information, uncertainty, entropy. Sources – BMS, Markov, RL The Source Coding Theorem