1 / 18

COMP 538 Reasoning and Decision under Uncertainty

COMP 538 Reasoning and Decision under Uncertainty. Introduction Readings: Pearl (1998, Chapter 1 Shafer and Pearl, Chapter 1. Objectives. Course objectives Course contents. Uncertainty. Uncertainty: the quality or state of being not clearly known. Uncertainty appears in many tasks

tanaya
Download Presentation

COMP 538 Reasoning and Decision under Uncertainty

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. COMP 538 Reasoning and Decision under Uncertainty Introduction Readings: Pearl (1998, Chapter 1 Shafer and Pearl, Chapter 1

  2. Objectives • Course objectives • Course contents

  3. Uncertainty • Uncertainty: the quality or state of being not clearly known. • Uncertainty appears in many tasks • Partial knowledge of the state of the world • Noisy observations • Phenomena that is not covered by our models • Inherent randomness

  4. Probability and Decision Theory • Well-known and well-understood framework for uncertainty • Clear semantics • Provides principled answers for: • Combining evidence • Predictive & Diagnostic reasoning • Incorporation of new evidence • Intuitive (at some level) to human experts • Can be learned

  5. Course Objectives • When applied to real-world problems, probability theory and decision theory suffer from • Complexity of model construction • Complexity of problem solving • This course covers methodologies developed recently in AI community for dealing with those complexity problems. • The methodologies combine ideas from several disciplines Artificial Intelligence, Machine Learning Decision Theory, Theory of Computer Science Statistics, Information Theory, Operations Research

  6. Complexity Problem of Applying Probability Theory • Example: • Patients in hospital are described by several attributes: • Background: age, gender, history of diseases, … • Symptoms: fever, blood pressure, headache, … • Diseases: pneumonia, heart attack, … • A joint probability distribution needs to assign a number to each combination of values of these attributes, exponential model size. • 20 attributes require 2020 ( roughly106 ) numbers • Real applications usually involve hundreds of attributes

  7. Complexity Problem of Applying Probability Theory • Because of the exponential model size problem, it was believed that probability theory is not practical for dealing with uncertainty in AI. • Alternative uncertainty calculi were introduced: uncertainty factors, non-monotonic logic, fuzzy logic, etc.

  8. MINVOLSET KINKEDTUBE PULMEMBOLUS INTUBATION VENTMACH DISCONNECT PAP SHUNT VENTLUNG VENITUBE PRESS MINOVL FIO2 VENTALV PVSAT ANAPHYLAXIS ARTCO2 EXPCO2 SAO2 TPR INSUFFANESTH HYPOVOLEMIA LVFAILURE CATECHOL LVEDVOLUME STROEVOLUME ERRCAUTER HR ERRBLOWOUTPUT HISTORY CO CVP PCWP HREKG HRSAT HRBP BP Complexity Problem of Applying Probability Theory • Bayesian networks alleviate the exponential model size problem. • Key idea: use conditional independence to factorize model into smaller parts. • Example: Alarm network 37 variables Model size: 237 Size of factored model: 509 Model construction and problem solving becomes possible

  9. Advantages of Bayesian Networks • Semantics • Probability theory provides the glue whereby the parts are combined, ensuring that the system as a whole is consistent. • Alternative approaches suffer from several semantic deficiencies (Pearl 1988, Chapter 1). • Model construction by expert • Appealing graphical interface allows experts to build model for highly interacting variables. • Model construction from data • Probability foundation allows model construction from data by well established statistical principles such as maximum likelihood estimation and Bayesian estimation.

  10. Fielded Applications • Expert systems • Medical diagnosis • Fault diagnosis (jet-engines, Windows 98) • Monitoring • Space shuttle engines (Vista project) • Freeway traffic • Sequence analysis and classification • Speech recognition • Biological sequences • Information access • Collaborative filtering • Information retrieval See tutorial by Breese and Koller (1997) and online resources for application samples.

  11. Course Content/Bayesian Networks • Concept and semantics of Bayesian networks • Inference: How to answer queries efficiently • Learning: How to learn/adapt Bayesian network models from data • Causal models: How to learn causality from statistical data Detailed discussion of those topics can take one whole course (DUKE). We will focus on the main ideas and skip the details so that we can study other related topics.

  12. Bayesian networks and classical multivariate models • Special cases of Bayesian networks: many of the classical multivariate models from • statistics, systems engineering, information theory, pattern recognition and statistical mechanics • Examples: mixture models, factor analysis, hidden Markov models, Kalman filters, Ising models. • Bayesian networks provide a way to view all of these models as instances of a common underlying formalism.

  13. Course Content/Special models • Latent class analysis: • Statistical method for finding subtypes of related cases. • With proper generalization, might provide a statistical foundation for Chinese medicine diagnosis. • Hidden Markov models: • A temporal model widely used in pattern recognition: handwriting recognition, speech recognition.

  14. Decision Making under Uncertainty • Typical scenario: whether to take umbrella • Decision theory provides basis for rational decision making: Maximum expected utility principle. • Decision analysis: applying of decision theory. • Suffers from exponential model size.

  15. Course Content/Simple Decision Making • Influence diagrams • Generalization of Bayesian network. • Alleviate the complexity problem of decision analysis • Topics: • Evaluation: How to compute optimal decisions in an influence diagram. • Value of information: Whether it is worthwhile to collect more information to reduce uncertainty.

  16. Sequential Decision Making under Uncertainty • Agent/robot needs to execute multiple actions in order to achieve a goal • Uncertainty originates from noisy sensor and inaccurate actuators/uncontrollable enviroment factors • What is the best way to reach goal with minimum cost/time?

  17. Course Content/Sequential Decision Making • Models: • Markov decision processes, consider only uncertainty in actuators/environment factors • Partially observable decision processes, consider uncertainty in sensors and actuators/environment factors. • Solution methods: • Value iteration • Policy iteration • Dealing with model complexity using dynamic Bayesian networks • Learning sequential decision models • Model-based reinforcement learning

  18. Course Content/Summary • Bayesian networks • Concept and semantics, inference, learning, causilty • Special models: Hidden Markov models, latent class analysis • Influence diagrams: • Evaluation, value of information • Markov decision processes and partially observable Markov decision processes • Solution methods: value iteration, policy iteration, dynamic Bayesian networks • Model-based reinforcement learning

More Related