1 / 44

Bayesian models as a tool for revealing inductive biases

Bayesian models as a tool for revealing inductive biases. Tom Griffiths University of California, Berkeley. Learning languages from utterances. blicket toma dax wug blicket wug. S  X Y X  {blicket,dax} Y  {toma, wug}. Learning functions from ( x , y ) pairs.

Download Presentation

Bayesian models as a tool for revealing inductive biases

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Bayesian models as a tool for revealing inductive biases Tom Griffiths University of California, Berkeley

  2. Learning languages from utterances blicket toma dax wug blicket wug S  X Y X  {blicket,dax} Y  {toma, wug} Learning functions from (x,y) pairs Learning categories from instances of their members Inductive problems

  3. Revealing inductive biases • Many problems in cognitive science can be formulated as problems of induction • learning languages, concepts, and causal relations • Such problems are not solvable without bias (e.g., Goodman, 1955; Kearns & Vazirani, 1994; Vapnik, 1995) • What biases guide human inductive inferences? How can computational models be used to investigate human inductive biases?

  4. Models and inductive biases • Transparent

  5. Bayesian models Reverend Thomas Bayes

  6. Likelihood Prior probability Posterior probability Sum over space of hypotheses Bayes’ theorem h: hypothesis d: data

  7. Three advantages of Bayesian models • Transparent identification of inductive biases through hypothesis space, prior, and likelihood • Opportunity to explore a range of biases expressed in terms that are natural to the problem at hand • Rational statistical inference provides an upper bound on human inferences from data

  8. Two examples Causal induction from small samples (Josh Tenenbaum, David Sobel, Alison Gopnik) Statistical learning and word segmentation (Sharon Goldwater, Mark Johnson)

  9. Two examples Causal induction from small samples (Josh Tenenbaum, David Sobel, Alison Gopnik) Statistical learning and word segmentation (Sharon Goldwater, Mark Johnson)

  10. See this? It’s a blicket machine. Blickets make it go. Let’s put this one on the machine. Oooh, it’s a blicket! Blicket detector (Dave Sobel, Alison Gopnik, and colleagues)

  11. “One cause” (Gopnik, Sobel, Schulz, & Glymour, 2001) • Two objects: A and B • Trial 1: A B on detector – detector active • Trial 2: B on detector – detector inactive • 4-year-olds judge whether each object is a blicket • A: a blicket (100% say yes) • B: almost certainly not a blicket (16% say yes) A Trial B Trial AB Trial A B

  12. Hypotheses: causal models A B A B A B A B E E E E Defines probability distribution over variables (for both observation, and intervention) (Pearl, 2000; Spirtes, Glymour, & Scheines, 1993)

  13. Prior and likelihood: causal theory • Prior probability an object is a blicket is q • defines a distribution over causal models • Detectors have a deterministic “activation law” • always activate if a blicket is on the detector • never activate otherwise (Tenenbaum & Griffiths, 2003; Griffiths, 2005)

  14. Prior and likelihood: causal theory P(h00) = (1 – q)2 P(h01) = (1 – q) q P(h10) = q(1 – q) P(h11) = q2 A B A B A B A B E E E E P(E=1 | A=0, B=0): 0 0 0 0 P(E=0 | A=0, B=0): 1 1 1 1 P(E=1 | A=1, B=0): 0 0 1 1 P(E=0 | A=1, B=0): 1 1 0 0 P(E=1 | A=0, B=1): 0 1 0 1 P(E=0 | A=0, B=1): 1 0 1 0 P(E=1 | A=1, B=1): 0 1 1 1 P(E=0 | A=1, B=1): 1 0 0 0

  15. Modeling “one cause” P(h00) = (1 – q)2 P(h01) = (1 – q) q P(h10) = q(1 – q) P(h11) = q2 A B A B A B A B E E E E P(E=1 | A=0, B=0): 0 0 0 0 P(E=0 | A=0, B=0): 1 1 1 1 P(E=1 | A=1, B=0): 0 0 1 1 P(E=0 | A=1, B=0): 1 1 0 0 P(E=1 | A=0, B=1): 0 1 0 1 P(E=0 | A=0, B=1): 1 0 1 0 P(E=1 | A=1, B=1): 0 1 1 1 P(E=0 | A=1, B=1): 1 0 0 0

  16. Modeling “one cause” P(h01) = (1 – q) q P(h10) = q(1 – q) P(h11) = q2 A B A B A B E E E P(E=1 | A=0, B=0): 0 0 0 0 P(E=0 | A=0, B=0): 1 1 1 1 P(E=1 | A=1, B=0): 0 0 1 1 P(E=0 | A=1, B=0): 1 1 0 0 P(E=1 | A=0, B=1): 0 1 0 1 P(E=0 | A=0, B=1): 1 0 1 0 P(E=1 | A=1, B=1): 0 1 1 1 P(E=0 | A=1, B=1): 1 0 0 0

  17. Modeling “one cause” P(h10) = q(1 – q) A B A is definitely a blicket B is definitely not a blicket E P(E=1 | A=0, B=0): 0 0 0 0 P(E=0 | A=0, B=0): 1 1 1 1 P(E=1 | A=1, B=0): 0 0 1 1 P(E=0 | A=1, B=0): 1 1 0 0 P(E=1 | A=0, B=1): 0 1 0 1 P(E=0 | A=0, B=1): 1 0 1 0 P(E=1 | A=1, B=1): 0 1 1 1 P(E=0 | A=1, B=1): 1 0 0 0

  18. “One cause” (Gopnik, Sobel, Schulz, & Glymour, 2001) • Two objects: A and B • Trial 1: A B on detector – detector active • Trial 2: B on detector – detector inactive • 4-year-olds judge whether each object is a blicket • A: a blicket (100% say yes) • B: almost certainly not a blicket (16% say yes) A Trial B Trial AB Trial A B

  19. Building on this analysis • Transparent

  20. Other physical systems From stick-ball machines… …to lemur colonies (Kushnir, Schulz, Gopnik, & Danks, 2003) (Griffiths, Baraff, & Tenenbaum, 2004) (Griffiths & Tenenbaum, 2007)

  21. Two examples Causal induction from small samples (Josh Tenenbaum, David Sobel, Alison Gopnik) Statistical learning and word segmentation (Sharon Goldwater, Mark Johnson)

  22. Bayesian segmentation • In the domain of segmentation, we have: • Data: unsegmented corpus (transcriptions). • Hypotheses: sequences of word tokens. • Optimal solution is the segmentation with highest prior probability = 1 if concatenating words forms corpus, = 0 otherwise. Encodes assumptions about the structure of language

  23. Brent (1999) • Describes a Bayesian unigram model for segmentation. • Prior favors solutions with fewer words, shorter words. • Problems with Brent’s system: • Learning algorithm is approximate (non-optimal). • Difficult to extend to incorporate bigram info.

  24. A new unigram model (Dirichlet process) Assume word wi is generated as follows: 1. Is wi a novel lexical item? Fewer word types = Higher probability

  25. A new unigram model (Dirichlet process) Assume word wi is generated as follows: 2.If novel, generate phonemic form x1…xm : If not, choose lexical identity of wi from previously occurring words: Shorter words = Higher probability Power law = Higher probability

  26. Unigram model: simulations • Same corpus as Brent (Bernstein-Ratner, 1987): • 9790 utterances of phonemically transcribed child-directed speech (19-23 months). • Average utterance length: 3.4 words. • Average word length: 2.9 phonemes. • Example input: yuwanttusiD6bUk lUkD*z6b7wIThIzh&t &nd6dOgi yuwanttulUk&tDIs ...

  27. Example results

  28. What happened? • Model assumes (falsely) that words have the same probability regardless of context. • Positing amalgams allows the model to capture word-to-word dependencies. P(D&t) = .024 P(D&t|WAts) = .46 P(D&t|tu) = .0019

  29. What about other unigram models? • Brent’s learning algorithm is insufficient to identify the optimal segmentation. • Our solution has higher probability under his model than his own solution does. • On randomly permuted corpus, our system achieves 96% accuracy; Brent gets 81%. • Formal analysis shows undersegmentation is the optimal solution for any (reasonable) unigram model.

  30. Bigram model (hierachical Dirichlet process) Assume word wi is generated as follows: • Is (wi-1,wi) a novel bigram? • If novel, generate wi using unigram model (almost). If not, choose lexical identity of wi from words previously occurring after wi-1.

  31. Example results

  32. Conclusions • Both adults and children are sensitive to the nature of mechanisms in using covariation • Both adults and children can use covariation to make inferences about the nature of mechanisms • Bayesian inference provides a formal framework for understanding how statistics and knowledge interact in making these inferences • how theories constrain hypotheses, and are learned

  33. A probabilistic mechanism? • Children in Gopnik et al. (2001) who said that B was a blicket had seen evidence that the detector was probabilistic • one block activated detector 5/6 times • Replace the deterministic “activation law”… • activate with p = 1- if a blicket is on the detector • never activate otherwise

  34. Deterministic vs. probabilistic Deterministic mechanism knowledge affects intepretation of contingency data Probability of being a blicket Probabilistic One cause

  35. A B Manipulating mechanisms I. Familiarization phase: Establish nature of mechanism same block At end of the test phase, adults judge the probability that each object is a blicket II. Test phase: one cause B Trial AB Trial

  36. Bayes People Manipulating mechanisms(n = 12 undergraduates per condition) Deterministic Probability of being a blicket Probabilistic One cause

  37. Bayes People Manipulating mechanisms (n = 12 undergraduates per condition) Deterministic Probability of being a blicket Probabilistic One control One cause Three control

  38. A B Acquiring mechanism knowledge I. Familiarization phase: Establish nature of mechanism same block At end of the test phase, adults judge the probability that each object is a blicket II. Test phase: one cause B Trial AB Trial

  39. Results with children • Tested 24 four-year-olds (mean age 54 months) • Instead of rating, yes or no response • Significant difference in one cause B responses • deterministic: 8% say yes • probabilistic: 79% say yes • No significant difference in one control trials • deterministic: 4% say yes • probabilistic: 21% say yes (Griffiths & Sobel, submitted)

  40. Comparison to previous results • Proposed boundaries are more accurate than Brent’s, but fewer proposals are made. • Result: word tokens are less accurate. Precision: #correct / #found= [= hits / (hits + false alarms)] Recall: #found / #true = [= hits / (hits + misses)] F-score: an average of precision and recall.

  41. Quantitative evaluation • Compared to unigram model, more boundaries are proposed, with no loss in accuracy: • Accuracy is higher than previous models:

  42. Two examples Causal induction from small samples (Josh Tenenbaum, David Sobel, Alison Gopnik) Statistical learning and word segmentation (Sharon Goldwater, Mark Johnson)

More Related