1 / 20

Knowledge Engineering for Bayesian Networks

Knowledge Engineering for Bayesian Networks. Probability theory for representing uncertainty. Assigns a numerical degree of belief between 0 and 1 to facts e.g. “it will rain today” is T/F. P(“it will rain today”) = 0.2 prior probability (unconditional) Conditional probability (Posterior)

marlo
Download Presentation

Knowledge Engineering for Bayesian Networks

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Knowledge Engineering for Bayesian Networks

  2. Probability theory for representing uncertainty • Assigns a numerical degree of belief between 0 and 1 to facts • e.g. “it will rain today” is T/F. • P(“it will rain today”) = 0.2 prior probability (unconditional) • Conditional probability (Posterior) • P(“it wil rain today” | “rain is forecast”) = 0.8 • Bayes’ Rule: P(H|E) = P(E|H) x P(H) P(E)

  3. Bayesian networks • Directed acyclic graphs • Nodes: random variables, • R: “it is raining”, discrete values T/F • T: temperature, continuous or discrete variable • C: color, discrete values {red, blue, green} • Arcs indicate dependencies (can have causal interpretation)

  4. X Flu Y Te Q Th Bayesian networks • Conditional Probability Distribution (CPD) • Associated with each variable • probability of each state given parent states “Jane has the flu” P(Flu=T) = 0.05 Models causal relationship “Jane has a high temp” P(Te=High|Flu=T) = 0.4 P(Te=High|Flu=F) = 0.01 Models possible sensor error “Thermometer temp reading” P(Th=High|Te=H) = 0.95 P(Th=High|Te=L) = 0.1

  5. Inference in Belief Networks • Main task of a belief network: Compute the conditional probability of a set of query variables given exact values for some evidence variables: P(query | evidence). • Belief networks are flexible enough so that any node can serve as either a query or an evidence variable.

  6. Flu Flu TB Flu Flu Y Te Te Te Y Te Th Th Th Diagnostic inference Causal inference Mixed inference Intercausal inference BN inference • Evidence: observation of specific state • Task: compute the posterior probabilities for query node(s) given evidence. Flu

  7. Building a BN • Choose a set of random variables Xi that describe the domain. • Missing variables may cause the BN unreliable.

  8. Building a BN • Choose a set of random variables Xi that describe the domain. • Order the variables into a list L • Start with an empty BN. • For each variable X in L do • Add X into the BN • Choose a minimal set of nodes already in the BN which satisfy the conditional dependence property with X • Make these nodes the parents of X. • Fill in the CPT for X.

  9. The Alarm Example • Mr. Holmes’ security alarm at home may be triggered by either burglar or earthquake. When the alarm sounds, his two nice neighbors, Mary and John, may call him. causal DAG

  10. The Alarm Example • Variable order: • Burglary • Earthquake • Alarm • JohnCalls • MaryCalls BN

  11. The Alarm Example • Variable order: • MaryCalls • JohnCalls • Alarm • Burglary • Earthquake BN

  12. The Alarm Example BN • Variable order: • MaryCalls • JohnCalls • Earthquake • Burglary • Alarm

  13. Weakness of BN • Hard to obtain JPD (joint probability distribution) • Relative Frequency Approach: counting outcomes of repeated experiments • Subjective Approach: an individual's personal judgment about whether a specific outcome is likely to occur. • Worst time complexity is NP-hard.

  14. BN software • Commerical packages: Netica, Hugin, Analytica (all with demo versions) • Free software: Smile, Genie, JavaBayes, … http://HTTP.CS.Berkeley.EDU/~murphyk/Bayes/bnsoft.html • Example running Netica software

  15. What’s Netica? • Netica is a powerful, easy-to-use, complete program for working with belief networksand influence diagrams. It has an intuitive and smooth user interface for drawing the networks, and the relationships between variables may be entered as individual probabilities, in the form of equations, or learned from data files.

  16. Netica Screen Shot Priori probabilities are needed for each variables. Netica will compute CPT (conditional probability table).

  17. Netica Screen Shot P(Jewelry = yes | Age < 30, Sex = Male) P(Fraud = yes | Jewelry = yes, Age < 30, Sex = male)

  18. Netica Screen Shot P(Fraud = yes | Gas = yes, Jewelry = yes, Age < 30, Sex = male) P(Fraud = yes | Gas = yes, Jewelry = yes, Age > 50, Sex = female)

  19. Extensions of BN • Weaker requirement in a DAG: Instead of I(X, NDX | PAX), ask I(X, NDX | MBX), where MBX is called Markov Blanket of X, which is the set of neighboring nodes: its parents (PAX), its children, and any other parents of X’s children. PAB = { H } MBB = { H, L, F } NDB = { L, X }

  20. Open Research Questions • Methodology for combining expert elicitation and automated methods • expert knowledge used to guide search • automated methods provide alternatives to be presented to experts • Evaluation measures and methods • may be domain depended • Improved tools to support elicitation • e.g. visualisation of d-separation • Industry adoption of BN technology

More Related