1 / 29

Bayesian models of human inference

Bayesian models of human inference. Josh Tenenbaum MIT. The Bayesian revolution in AI . Principled and effective solutions for inductive inference from ambiguous data: Vision Robotics Machine learning Expert systems / reasoning Natural language processing

vine
Download Presentation

Bayesian models of human inference

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Bayesian models of human inference Josh Tenenbaum MIT

  2. The Bayesian revolution in AI • Principled and effective solutions for inductive inference from ambiguous data: • Vision • Robotics • Machine learning • Expert systems / reasoning • Natural language processing • Standard view in AI: no necessary connection to how the human brain solves these problems. • Heuristics & Biases program in the background (“We know people aren’t Bayesian, but…”).

  3. Bayesian models of cognition Visual perception [Weiss, Simoncelli, Adelson, Richards, Freeman, Feldman, Kersten, Knill, Maloney, Olshausen, Jacobs, Pouget, ...] Language acquisition and processing [Brent, de Marken, Niyogi, Klein, Manning, Jurafsky, Keller, Levy, Hale, Johnson, Griffiths, Perfors, Tenenbaum, …] Motor learning and motor control [Ghahramani, Jordan, Wolpert, Kording, Kawato, Doya, Todorov, Shadmehr,…] Associative learning [Dayan, Daw, Kakade, Courville, Touretzky, Kruschke, …] Memory [Anderson, Schooler, Shiffrin, Steyvers, Griffiths, McClelland, …] Attention [Mozer, Huber, Torralba, Oliva, Geisler, Yu, Itti, Baldi, …] Categorization and concept learning [Anderson, Nosfosky, Rehder, Navarro, Griffiths, Feldman, Tenenbaum, Rosseel, Goodman, Kemp, Mansinghka, …] Reasoning [Chater, Oaksford, Sloman, McKenzie, Heit, Tenenbaum, Kemp, …] Causal inference [Waldmann, Sloman, Steyvers, Griffiths, Tenenbaum, Yuille, …] Decision making and theory of mind [Lee, Stankiewicz, Rao, Baker, Goodman, Tenenbaum, …]

  4. How to meet up with mainstream JDM research (i.e., heuristics & biases)? • How to reconcile apparently contradictory messages of H&B and Bayesian models? Are people Bayesian or aren’t they? When are they, when aren’t they, and why? • How to integrate the H&B and Bayesian research approaches?

  5. When are people Bayesian, and why? • Low level hypothesis (Shiffrin, Maloney, etc.) • People are Bayesian in low-level input or output processes that have a long evolutionary history shared with other species, e.g. vision, motor control, memory retrieval.

  6. When are people Bayesian, and why? • Higher-level cognition can be Bayesian when information is presented in formats that we have evolved to process, and that support simple heuristic algorithms, e.g., base-rate neglect disappears with “natural frequencies”. • Low level hypothesis (Shiffrin, Maloney, etc.) • Information format hypothesis (Gigerenzer) Explicit probabilities Natural frequencies

  7. When are people Bayesian, and why? • Low level hypothesis (Shiffrin, Maloney, etc.) • Information format hypothesis (Gigerenzer) • Core capacities hypothesis • Bayes can illuminate core human cognitive capacities for inductive inference – learning words and concepts, projecting properties of objects, causal inference, or action understanding: problems we solve effortlessly, unconsciously, and successfully in natural contexts, which any five-year-old solves better than any animal or computer.

  8. A B E When are people Bayesian, and why? • Low level hypothesis (Shiffrin, Maloney, etc.) • Information format hypothesis (Gigerenzer) • Core capacities hypothesis A B A Causal induction B A B B A Trial AB Trial ? ? (Sobel, Griffiths, Tenenbaum, & Gopnik)

  9. When are people Bayesian, and why? • Low level hypothesis (Shiffrin, Maloney, etc.) • Information format hypothesis (Gigerenzer) • Core capacities hypothesis Word learning Hypothesis space Data (Tenenbaum & Xu)

  10. When are people Bayesian, and why? • Low level hypothesis (Shiffrin, Maloney, etc.) • Information format hypothesis (Gigerenzer) • Core capacities hypothesis • Bayes can illuminate core human cognitive capacities for inductive inference – learning words and concepts, projecting properties of objects, causal inference, or action understanding: problems we solve effortlessly, unconsciously, and successfully in natural contexts, which a five-year-old solves better than any animal or computer. • The mind is not good at explicit Bayesian reasoning about verbally or symbolically presented statistics, unless core capacities can be engaged.

  11. When are people Bayesian, and why? • Low level hypothesis (Shiffrin, Maloney, etc.) • Information format hypothesis (Gigerenzer) • Core capacities hypothesis Correct Statistical version of Diagnosis problem Causal version of Diagnosis problem Base-rate neglect (Krynski & Tenenbaum)

  12. How to meet up with mainstream JDM research (i.e., heuristics & biases)? • How to reconcile apparently contradictory messages of H&B and Bayesian models? Are people Bayesian or aren’t they? When are they, when aren’t they, and why? • How to integrate the H&B and Bayesian research approaches?

  13. Reverse engineering • Goal is to reverse-engineer human inference. • A computational understanding of how the mind works and why it works it does. • Even for core inferential capacities, we are likely to observe behavior that deviates from any ideal Bayesian analysis. • These deviations are likely to be informative about how the mind works.

  14. Analogy to visual illusions (Adelson) (Shepard) • Highlight the problems the visual system is designed to solve: inferring world structure from images, not judging properties of the images themselves. • Reveal the implicit visual system’s implicit assumptions about the physical world and the processes of image formation that are needed to solve these problems.

  15. How do we interpret deviations from a Bayesian analysis? • H&B: People aren’t Bayesian, but use some other means of inference. • Base-rate neglect: representativeness heuristic • Recency bias: availability heuristic • Order of evidence effects: anchoring and adjustment • … • Not so compelling as reverse engineering. • What engineer would want to design a system based on “representativeness”, without knowing how it is computed, why it is computed that way, what problem it attempts to solve, when it works, or how its accuracy and efficiency compares to some ideal computation or other heuristics.

  16. How do we interpret deviations from a Bayesian analysis? Multiple levels of analysis (Marr) • Computational theory • What is the goal of the computation – the outputs and available inputs? What is the logic by which the inference can be performed? What constraints (prior knowledge) do people assume to make the solution well-posed? • Representation and algorithm • How is the information represented? How is the computation carried out algorithmically, approximating the ideal computational theory with realistic time & space resources? • Hardware implementation

  17. How do we interpret deviations from a Bayesian analysis? Multiple levels of analysis (Marr) • Computational theory • What is the goal of the computation – the outputs and available inputs? What is the logic by which the inference can be performed? What constraints (prior knowledge) do people assume to make the solution well-posed? • Representation and algorithm • How is the information represented? How is the computation carried out algorithmically, approximating the ideal computational theory with realistic time & space resources? • Hardware implementation Bayes

  18. Different philosophies • H&B • One canonical Bayesian analysis of any given task, and we know what it is. • Ideal Bayesian solution can be computed. • The question “Are people Bayesian?” is empirically meaningful on any given task. • Bayes+Marr • Many possible Bayesian analyses of any given task, and we need to discover which best characterize cognition. • Ideal Bayesian solution can only be approximately computed. • The question “Are people Bayesian?” is not an empirical one, at least not for an individual task. Bayes is a framework-level assumption, like distributed representations in connectionism or condition-action rules in ACT-R.

  19. How do we interpret deviations from a Bayesian analysis? Multiple levels of analysis (Marr) • Computational theory • What is the goal of the computation – the outputs and available inputs? What is the logic by which the inference can be performed? What constraints (prior knowledge) do people assume to make the solution well-posed? • Representation and algorithm • How is the information represented? How is the computation carried out algorithmically, approximating the ideal computational theory with realistic time & space resources? • Hardware implementation

  20. The centrality of causal inference • In visual perception: • Judge P(scene|image features) rather than P(image features|scene) or P(image features|other image features). • Coin–flipping: Which sequence is more likely to come from flipping a fair coin, HHTHT or HHHHH? • Coincidences: How likely that 2 people in a random party of 25 have the same birthday? 3 in a party of 10? (Griffiths & Tenenbaum)

  21. Judgments of randomness: Judgments of coincidence: Rational measure of evidential support: (Griffiths & Tenenbaum)

  22. How do we interpret deviations from a Bayesian analysis? Multiple levels of analysis (Marr) • Computational theory • What is the goal of the computation – the outputs and available inputs? What is the logic by which the inference can be performed? What constraints (prior knowledge) do people assume to make the solution well-posed? • Representation and algorithm • How is the information represented? How is the computation carried out algorithmically, approximating the ideal computational theory with realistic time & space resources? • Hardware implementation

  23. Assuming the world is simple • In visual perception: • “Slow and smooth” prior on visual motion • Causal induction: • P(blicket) = 1/6, “Activation law” P(A is a blicket|data) = 1 P(B is a blicket|data) ~ 1/6 P(A is a blicket|data) ~ 3/4 P(B is a blicket|data) ~ 1/4 A B A B A A B B A Trial AB Trial C A B A C B C A B AC Trial AB Trial

  24. Gorillas have T9 cells. Seals have T9 cells. Horses have T9 cells. Recognizing the world is complex • In visual perception: • Need uncertainty about coherence ratio and velocity of coherent motion. (Lu & Yuille) • Property induction: • Properties should be distributed stochastically over tree structure, not just focused on single branches. r = 0.50 Bayes: single branch prior (Kemp & Tenenbaum)

  25. Gorillas have T9 cells. Seals have T9 cells. Horses have T9 cells. Recognizing the world is complex • In visual perception: • Need uncertainty about coherence ratio and velocity of coherent motion. (Lu & Yuille) • Property induction: • Properties should be distributed stochastically over tree structure, not just focused on single branches. r = 0.92 Bayes: “mutation” prior (Kemp & Tenenbaum)

  26. “has T9 hormones” “can bite through wire” “is found near Minneapolis” “carry E. Spirus bacteria” (Kemp & Tenenbaum)

  27. How do we interpret deviations from a Bayesian analysis? Multiple levels of analysis (Marr) • Computational theory • What is the goal of the computation – the outputs and available inputs? What is the logic by which the inference can be performed? What constraints (prior knowledge) do people assume to make the solution well-posed? • Representation and algorithm • How is the information represented? How is the computation carried out algorithmically, approximating the ideal computational theory with realistic time & space resources? • Hardware implementation

  28. Sampling-based approximate inference • In visual perception: • Temporal dynamics of bi-stability due to fast sampling-based approximation of a bimodal posterior (Schrater & Sundareswara). • Order effects in category learning • Particle filter (sequential Monte Carlo), an online approximate inference algorithm assuming stationarity. • Probability matching in classification decisions • Sampling-based approximations with guarantees of near optimal generalization performance. (Griffiths et al., Goodman et al.)

  29. Conclusions • “Are people Bayesian?”, “When are they Bayesian?” • Maybe not the most interesting questions in the long run…. • What is the best way to reverse engineer cognition at multiple levels of analysis? Assuming core inductive capacities are approximately Bayesian at the computational-theory level offers several benefits: • Explanatory power: why does cognition work? • Fewer degrees of freedom in modeling • A bridge to state-of-the-art AI and machine learning • Tools to study the big questions: What are the goals of cognition? What does the mind know about the world? How is that knowledge represented? What are the processing mechanisms and why do they work as they do?

More Related