logo

Problems_ Chapter 3


Problems _ Chapter 3 3.1 A discrete source generates three independent symbols A, B and C with probabilities 0.9, 0.08 and 0.02 respectively. a) Determine the entropy of the source. b) Determine the redundancy of the source. 3.2 a) Consider a source having an M = 3 symbol alphabet where p(x1) = ½, p(x2) = p(x3) = ¼ and symbols are statistically independent. Calculate the information conveyed by the receipt of the symbol x1. Repeat for x2 and x3. b) Consider a source whose, statistically independent, symbols consist of all possible binary sequences of length k. Assume all symbols are equiprobable. How much information is conveyed on receipt of any symbol. c) Dermine the information conveyed by the specific message x1x3x2x4 when it emanates from the following statistically independent symbol source: M = 4, p(x1) = 1/2, p(x2) = 1/4, p(x3) = p(x4) = 1/8. 3.3 Calculate the loss information due to noise, per transmitted digit, if a random binary signal is transmitted through a channel, which adds zero mean Gaussian noise, with an average signal-to-noise ratio of: a) 0 dB b) 5 dB c) 10 dB 3.4 An information source contains 100 different, statistically independent, equiprobable symbols. Find the maximum code efficiency, if, for transmission, all the symbols are represented by binary code words of equal length. 3.5 a) Apply Huffman’s algorithm to deduce an optimal code for transmitting the source defined in problem 3.2a over a binary channel. Is your code unique? b) Determine the efficiency of the code devised in part (a). c) Construct another code for the source of part (a) and assign equal length binary words irrespective of the occurrence probability of the symbols. Calculate the efficiency of this source. 3.6 An input alphabet (a keyboard on a word processor) consists of 100 characters. a) If the keystrokes are encoded by a fixed-length code, determine the required number of bits for the encoding b) We make the assumption that 10 of the keystrokes are equally likely and that each occurs with probability of 0.05. We also assume that the remaining 90 keystrokes are equally likely. Determine the average number of bits required to encode this alphabet using a variable-length Huffman code. 3.7 Encode the specific message AABACB using: a) The 7-bit ASCII. b) The adaptive Huffman code. c) Compare the number of bits in encoded message in part (a) and part (b). Give your comments.
DMCA.com Protection Status Copyright by webtailieu.net