1 / 87

Universal Modeling: Introduction to ‘Modern’ MDL

Universal Modeling: Introduction to ‘Modern’ MDL. Peter Gr ü nwald CWI Amsterdam www.cwi.nl/~pdg. Further Reading: P. Gr ü nwald. The Minimum Description Length Principle , MIT Press, 2007. TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A A A.

haracha
Download Presentation

Universal Modeling: Introduction to ‘Modern’ MDL

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Universal Modeling:Introduction to ‘Modern’ MDL Peter Grünwald CWI Amsterdam www.cwi.nl/~pdg Further Reading: P. Grünwald. The Minimum Description Length Principle, MIT Press, 2007. TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAAA

  2. Overview • Introduction • Probability and Code Length • Universal Models • MDL Model Selection • Interpretation • New Developments

  3. Overview • Introduction • Probability and Code Length • Universal Models • MDL Model Selection • Interpretation • New Developments

  4. Minimum Description Length Principle Rissanen 1978, 1987, 1996, Barron, Rissanen and Yu 1998 • ‘MDL’ is a method for inductive inference, • in particular developed and suited for model selection problems • but can do prediction/estimation as well

  5. Minimum Description Length Principle • MDL is based on the correspondence between ‘regularity’ and ‘compression’: • The more you are able to compress a sequence of data, the more regularity you have detected in the data • Example:

  6. Minimum Description Length Principle • MDL is based on the correspondence between ‘regularity’ and ‘compression’: • The more you are able to compress a sequence of data, the more regularity you have detected in the data… • …and thus the more you have learned from the data: • ‘inductive inference’ as trying to find regularities in data (and using those to make predictions of future data)

  7. Model Selection • Given data and ‘models’ • , • which model best explains the data ? • Need to take into account • Error (minus Goodness-of-fit) • Complexity of models • Examples • Variable (order) selection in regression • Selection of order in (hidden) Markov Models

  8. ‘Modern’ MDL? Algorithmic, ‘ideal’ MDL (Li and Vitányi ’97) MML (Wallace ‘68 (!), ‘87) Kinds of MDL 2-part code MDL (Rissanen ’78, ’83) Universal model based MDL (Rissanen ’96, Barron, Rissanen, Yu ‘98, Grünwald ‘07)

  9. Modern MDL! Algorithmic, ‘ideal’ MDL (Li and Vitányi ’97) MML (Wallace ‘68 (!), ‘87) Kinds of MDL 2-part code MDL (Rissanen ’78, ’83) Universal model based MDL (Rissanen ’96, Barron, Rissanen, Yu ‘98, Grünwald ‘07)

  10. Overview • Introduction • Probability and Code Length • Universal Models • MDL Model Selection • Interpretation • Predictive MDL Estimation

  11. Codes • (countable) ‘data alphabet’ • A (uniquely decodable) code is a one-to-one map from to . • denotes the length (in bits) needed to describe .

  12. Code Length & Probability • Let be a probability distribution. Since only few can have ‘large’ probability • Let be a code for . . Since the fraction of sequences that can be compressed by more than bits is less than , only very few symbols can have small code length. • This suggests an analogy!

  13. Code Lengths ‘are’ probabilities… • Let be a (uniquely decodable) code over countable set . Then there exists a (possibly defective) probability distribution such that • is a ‘proper’ probability distribution iff the code is ‘complete’. (follows from Kraft-McMillan inequality)

  14. …and probabilities ‘are’ code lengths! • Let be a probability distribution over countable set . Then there exists a code for such that

  15. The Most Important Slide! There is a 1-1 correspondence between probability distributions and code length functions, such that small probabilities correspond to large code lengths and vice versa:

  16. The Most Important Slide! There is a 1-1 correspondence between probability distributions and code length functions, such that small probabilities correspond to large code lengths and vice versa: Example: is 1st 0rder Markov Chain – if fits data well (regularities in data are well-captured by ) , the code based on compresses much.

  17. Remarks • In this correspondence, we do not assume that data are sampled from a probability distribution! • Extend correspondence to continuous sample space through discretization; may stand for density • Distributions and codes over sequences of outcomes: still max. 1 bit round-off error • Neglect difference and identify code length functions and probability mass functions

  18. Overview • Introduction • Probability and Code Length • Universal Models • MDL Model Selection • Interpretation • Predictive MDL Estimation

  19. Universal Codes • : set of code (length function)s available to encode data • Suppose we think that one of the code(length function)s in allows for substantial compression of • GOAL: encode using minimum number of bits!

  20. Universal Codes • Simply encoding using the that minimizes code length does not work (encoding cannot be decoded) • But there exist codes which, for any sequence are ‘almost’ as good as • These are called ‘universal codes’ for

  21. Universal Codes • Example: finite • There exists a code such that for some constant , for all , ,all : • In particular, • Note that does not depend on , while typically, grows linearly in

  22. Universal Models • Let be a probabilistic model, i.e. a family (set) of probability distributions • Assume finite: • There exists a code such that for all : • Hence, exists distribution such that • i.e. • is a ‘universal model’ (distribution) for

  23. Terminology • Statistics: • Model = family of distributions • Information theory: • Model = single distribution • Model class = family of distributions • Universal model is a single distribution acting as a representative of/defined relative to a set of distributions

  24. Bayesian Mixtures are universal models • Let be a prior over . The Bayesian marginal likelihood is defined as:

  25. Bayesian Mixtures are universal models • Let be a prior over . The Bayesian marginal likelihood is defined as: • This is a universal model, since

  26. 2-part MDL code is a universal model (code) • The ML (maximum likelihood) distribution is the achieving • Code by first coding , then coding ‘with the help of’ :

  27. 2-part vs. Bayes universal models • Bayes’ mixture strictly ‘better’ universal model in that it assigns larger probability (shorter code length) to outcomes. • What does ‘better’ really mean? • What prior leads to short code lengths?

  28. Optimal Universal Model • Look for such that regret • is small no matter what are; i.e. look for

  29. Optimal Universal Model - II • is achieved for Normalized Maximum Likelihood (NML) distribution (Shtarkov 1987):

  30. MDL Model Selection • Suppose we are given data • We want to select between models and as explanations for the data. MDL tells us to pick the for which the associated optimal universal model assigns the largest probability to the data:

  31. MDL Model Selection • Select minimizing , i.e. minimizing error (= minus fit) term complexity term ( )

  32. Four Interpretations • Compression interpretation • Select model that compresses data most, treating all distributions within model on equal footing; detects most (non-spurious) regularity in data • Counting/Geometric interpretation • Bayesian interpretation • Predictive interpretation

  33. Counting Interpretation of MDL • Select minimizing , i.e. minimizing error (= minus fit) term complexity term ( ) Something like ‘total fit’ model gives to data Log ‘effective’ number of distributions

  34. Counting Interpretation of MDL number of distributions total amount of confusion

  35. Counting Interpretation of MDL • Select minimizing , i.e. minimizing error (= minus fit) term complexity term ( ) Something like ‘total fit’ model gives to data Log number of ‘distinguishable’ distributions

  36. Parametric Model Classes • Under regularity conditions: • Here: Number of free parameters in Fisher information matrix at Goes to 0 as

  37. Geometric Interpretation of MDL • Under regularity conditions: • Compare BIC (Schwartz ’78), old ‘MDL Criterion’ (Rissanen ’78): select minimizing:

  38. Geometric Interpretation of MDL • Under regularity conditions: nr of parameters Curvature term Error Term Complexity Terms

  39. Bayesian Model Selection vs. MDL • Under regularity conditions: • Under regularity conditions: • Always within O(1) ; hence, for large enough n, Bayes and MDL select the same model

  40. Bayesian Model Selection vs. MDL • Under regularity conditions: • Under regularity conditions: • If we take Jeffreys-Bernardo prior, • within o(1): Bayes and NML become indistinguishable

  41. Bayes and MDL, remarks • Jeffreys’ prior was proposed as a ‘non-informative Bayesian prior’ by Jeffreys in 1939 • Jeffreys’ prior is uniform prior not on parameter space but on the space of distributions with the ‘natural metric’ that measures distances between distributions by how distinguishable they are • (but MDL is not Bayes! • e.g., MDL is immune to Diaconis-Freedman nonparametric inconsistency results)

  42. Further topics • Predictive interpretation (MDL as an automatic cross-validation like procedure) • Comparing infinitely many models • Predictive MDL Estimation • Frequentist justification

  43. Predictive Interpretation • Interpret as ‘loss’ incurred when predicting using while actual outcome was • Bayesian codelength can be rewritten as • accumulated log-loss prediction error • Here is the • Bayesian predictive distribution (posterior mixture)

  44. Predictive Interpretation, II • Idea (Dawid/Rissanen): for large , Bayesian predictive distribution resembles ML distribution more and more; therefore, may try to approximate by • or more generally by • for any ‘likelihood-based estimator’

  45. Predictive Interpretation, III • It turns out that (under regularity conditions) • Hence, ‘predictive code’ is a universal code • MDL model selection picks the model such that sequential prediction of the future given the past within the observed data leads to lowest accumulated sequential prediction error.

  46. Predictive Interpretation, IV • MDL can be cast in terms of prequential validation (Dawid ’84) • similar to leave-one-out cross-validation • essential difference: in MDL/prequential validation, if value of is used in prediction of , then value of not used in prediction of • If number of models under consideration is finite and constant, but , then • Prequential validation/MDL like BIC • Leave-One-Out CV like AIC

  47. Comparing infinitely many models • Select minimizing • i.e. minimizing • Reason: whole procedure should be interpretable as minimizing codelength for data. • We implicitly used uniform code to encode before.

  48. Comparing infinitely many models • Better not use two-part code for parameters • NML, Bayes give smaller regret (relative code-lengths) • We are forced to use two-part code for encoding model index • Because we want to select a model, we explicity have to encode it • Note: complexity of models not due to model index!

  49. Overview • Introduction • Probability and Code Length • Universal Models • MDL Model Selection • Interpretation • Overview of New Developments • Predictive MDL Estimation/Justification

  50. New Developments • Efficient Calculation of NML for some model classes (Myllymaki, 11.00 today) • What if NML distribution is undefined? • luckiness principle (Luckiness NML, Conditional NML) • Sequential NML (Silander, 14.30 today) • Nonparametrics (Seeger, 16.30 today) • Inherent improvement by adopting different notion of universality; solving AIC-BIC dilemma • (De Rooij, 11.30, Grunwald, 9.30 tomorrow)

More Related