1 / 10

Information Theory

Information Theory. Eighteenth Meeting. A Communication Model. Messages are produced by a source transmitted over a channel to the destination. encoded for transmission via the communication channel, then decoded at the destination. Codes. source symbols

bblackwell
Download Presentation

Information Theory

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Information Theory Eighteenth Meeting

  2. A Communication Model • Messages are • produced by a source • transmitted over a channel to the destination. • encoded for transmission via the communication channel, then • decoded at the destination.

  3. Codes • source symbols • A fixed set of source symbols (source alphabet) • code words are a representation for transmission. • A Sequence of binary digits • sometimes described as the code alphabet. • For a binary code, the code alphabet consists of only the symbols 0 and 1. • radix • The number of symbols in the code alphabet • binary code has radix 2. • Morse code uses dot, dash and space, has radix 3. • efficient code • minimize the number of binary digits needed.

  4. Information Probability • Efficiency of a code to be quantified. • The information conveyed by a message depends on the probability of receiving that particular message. • The information, I, gained by receiving a message of probability P is given by: • I = -log (P) • Example • Consecutive video frames typically have very similar content. • two consecutive frames is considerably less than the sum of the information in each individual frame. 000101010101010101010 000101010101010101010

  5. Information Entroby • Memoryless source • Two independent message produced by the source • I = −log (P1) + (−log (P2)) = −log (P1P2) • Entropy of the source (H). • probabilities of all the possible source symbols are known, • the average information for the source can be found. • the source can generate n different symbols and the ith symbol has probability Pi, then the entropy H is given by: • Representing the average amount of information the source provides. • The source entropy H is the minimum achievable value for the average length L of the code words: • L ≥ H • The average code word length L is given by: • The efficiency E :

  6. Variable length Carefully design Uniquely and instantaneously decodable. Example: four source symbols are represented by the binary sequences 0, 01, 011 and 111. message is: 00101101110 working from the right to the left. 0, 01, 011, 0, 111, 0 Coding Tree

  7. Huffman Code • source symbols with the largest probabilities are allocated systematically to the shortest code words • Original Source: are ordered according to their probabilities, • First reduction: the two lowest probabilities (0.1 and 0.02) • Second reduction: probabilities 0.12 and 0.13 are combined to give 0.25, • If necessary, the resulting values are re-ordered to ensure that the values in a column are in descending order. • This process is repeated until a probability of 1.0 is reached.

  8. Channel Coding • Assumptions • that the error rate is low, • that errors occur independently of each other, and • that there is a negligible probability of several errors occurring together.

  9. Rectangular Code

  10. Hamming Code • 1 = even parity for 357 • 2 = even parity for 367 • 4 = even parity for 567 1011

More Related