Due 22 September, 2016

Give reasons for all answers.

- Suppose you want to encode a source that produces characters
with the following distribution:

P(a_{1}) = 0.3, P(a_{2}) = 0.2, P(a_{3}) = 0.2, P(a_{4}) = 0.1, P(a_{5}) = 0.08, P(a_{6}) = 0.06, P(a_{7}) = 0.04, P(a_{8}) = 0.02

Design binary, ternary, and quaternary Huffman codes for this source and compare the average lengths to the entropy. Remember to use the appropriate base for the logarithms when calculating the entropies. - Design a Tunstall code for a source that produces independent
characters with probabilities
P(a
_{1}) = 0.5, P(a_{2}) = 0.3, P(a_{3}) = 0.2 where the codewords are blocks of two ternary digits. - In class we discussed how to encode an arbitrary integer using the ordinary binary representation of n preceded by a unary value giving the number of bits needed to represent n. For example, 1 is encoded as 01, 2 is encoded as 1010, 5 is encoded as 110101, etc. Find a probability distribution on the positive integers for which this is an optimal code. Try to find a simple expression for the probabilities.
- Use the Kraft-McMillan inequality to show that there is no encoding of the positive integers that uses fewer than log n bits to represent n for every n.