1 / 7

Lecture 3. Symmetry of Information

Lecture 3. Symmetry of Information. In Shannon information theory, the symmetry of information is well known. The same phenomenon is also in Kolmogorov complexity although the proof is totally different as well as the meaning.

annick
Download Presentation

Lecture 3. Symmetry of Information

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Lecture 3. Symmetry of Information • In Shannon information theory, the symmetry of information is well known. The same phenomenon is also in Kolmogorov complexity although the proof is totally different as well as the meaning. • The term C(x)-C(x|y) is known as “the information y knows about x”. This information is symmetric was first proved by Levin and Kolmogorov (in Zvonkin-Levin, Russ. Math Surv, 1970) • Today, we prove the important Symmetry of Information Theorem: Theorem. C(x)-C(x|y)=C(y)-C(y|x), modulo a log-term.

  2. Symmetry in our lives • Biological • Crystal • Greek architecture • M.C. Escher’s art

  3. Symmetry of Information Theorem Theorem. C(x)-C(x|y)=C(y)-C(y|x), modulo a log-term. Proof. Essentially we will prove: C(xy)=C(y|x)+C(x) + O(logC(xy)). (Since then C(xy)=C(x|y)+C(y) + O(logC(xy)). Theorem follows). (≤). It is trivial to see that C(xy)≤C(y|x)+C(x) + O(logC(xy)) is ture. (≥). We now need to prove C(xy)≥C(y|x)+C(x) + O(logC(xy)).

  4. The idea ---To show: C(xy)≥C(y|x)+C(x) + O(logC(xy)) A = {(u,z)|C(u,z)<C(x,y)} |A| ≤ 2c(x,y) Ax={z: C(x,z) ≤ C(x,y)} Ax’’ • This forms a partition of A • Ax must have >2e elements • otherwise C(y|x) too small. • There cannot be more than • |A| / 2e such blocks. Thus • C(x) is small. • e = C(xy)-C(x) – clogC(xy) Ax Ax’

  5. Proving: C(xy) ≥C(y|x)+C(x) + O(logC(xy)). Assume to the contrary: for each c, there are x and y s.t. C(xy)<C(y|x)+C(x) + clog C(xy) (1) Let A={(u,z): C(u,z) ≤ C(x,y)}. Given C(x,y), the set A can be recursively enumerated. Let Ax={z: C(x,z) ≤ C(x,y)}. Given C(x,y) and x, we have a simple algorithm to recursively enumerate Ax. One can describe y, given x, using its index in the enumeration of Ax and C(x,y). Hence C(y|x) ≤ log |Ax| + 2logC(x,y) + O(1) (2) By (1) and (2), for each c, there are x, y s.t. |Ax|>2e, where e=C(x,y)-C(x)-clogC(x,y). But now, we obtain a too short description for x as follows. Given C(x,y) and e, we can recursively enumerate the strings u which are candidates for x by satisfying Au={z: C(u,z) ≤ C(x,y)}, and 2e <|Au|. (3) Denote the set of such u by U. Clearly, x ε U. Also {<u,z> : u ε U & z ε Au} ⊆A(4) The number of elements in A cannot exceed the available number of programs that are short enough to satisfy its definition: |A| ≤ 2C(x,y)+O(1) (5) Using (3)(4)(5), |U| < |A|/2e ≤ 2C(x,y)+O(1) / 2e Hence we can reconstruct x from C(x,y), e, and the index of x in the enumeration of U. Therefore C(x) < 2logC(x,y) + 2loge + C(x,y) – e +O(1) Substituting e=C(x,y)-C(x)+clogC(x,y) yields C(x)<C(x), for large c, contradiction. QED

  6. Kamae Theorem For each natural number m, there is a string x such that for all but finitely many strings y, C(x) – C(x|y) ≥ m That is: there exist finite objects such that almost all finite objects contain a large amount of information about them.

  7. Relations to Shannon Information theory • Shannon entropy H(X) = - ∑ P(X=x)logP(X=x) Interpretation: H(X) bits are sufficient, on the average, to describe the outcome x. • C(x), in Kolmogorov complexity, is the minimum description (smallest program) for one fixed x. It is a beautiful fact that these two notions can be shown to converge to the same thing, Theorem 2.8.1. • The symmetry of information is another example that the two concepts coincide.

More Related