1 / 36

Learning Models of Relational Stochastic Processes

Learning Models of Relational Stochastic Processes. Sumit Sanghai. P2. P1. P3. V1. A1. A2. Motivation. Features of real-world domains Multiple classes, objects, relations. P2. ?. P1. P3. V1. A1. A2. ?. ?. Motivation. Features of real-world domains

edgonzalez
Download Presentation

Learning Models of Relational Stochastic Processes

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Learning Models of Relational Stochastic Processes Sumit Sanghai

  2. P2 P1 P3 V1 A1 A2 Motivation • Features of real-world domains • Multiple classes, objects, relations

  3. P2 ? P1 P3 V1 A1 A2 ? ? Motivation • Features of real-world domains • Multiple classes, objects, relations • Uncertainty

  4. Motivation • Features of real-world domains • Multiple classes, objects, relations • Uncertainty • Changes with time P5 P2 P6 P1 P3 V1 P4 A3 A1 A2

  5. Relational Stochastic Processes • Features • Multiple classes, objects, relations • Uncertainty • Change over time • Examples: Social networks, molecular biology, user activity modeling, web, plan recognition, … • Growth inherent or due to explicit actions • Most large datasets are gathered over time • Explore dependencies over time • Predict future

  6. Manufacturing Process

  7. Paint(A, blue) Manufacturing Process

  8. Paint(A, blue) Bolt(B, C) Manufacturing Process

  9. A.col A.col P t t + 1 red red 0.1 red blue 0.9 blue blue 0.95 blue red 0.05 Manufacturing Process Paint(A, blue) Bolt(B, C)

  10. Why are they different? • Modeling object, relationships modification, creation and deletion • Modeling actions (preconditions/effects), activities, plans • Can’t just throw ``time’’ into the mix • Summarizing object information • Learning can be made easier by concentrating on temporal dependencies • Sophisticated inference techniques like particle filtering may be applicable

  11. Outline • Background: Dynamic Bayes Nets • Dynamic Probabilistic Relational Models • Inference in DPRMs • Learning with Dynamic Markov Logic Nets • Future Work

  12. X1t+1 X1t X2t X2t+1 Yt+1 Yt t t+1 Dynamic Bayesian Networks • DBNs model change in uncertain variables over time • Each time slice consists of state/observation variables • Bayesian network models dependency of current on previous time slice(s) • At each node a conditional model (CPT, logistic regression, etc.)

  13. Inference and learning in DBNs • Inference • All techniques from BNs are used • Special techniques like Particle Filtering, Boyen-Koller, Factored Frontier, etc. can be used for state monitoring • Learning • Problem exactly similar to BNs • Structural EM used in case of missing data • Needs a fast inference algorithm

  14. Particle Filtering in DBNs • Task: State monitoring • Particle Filter • Samples represent state distribution at time t • Generate samples for t+1 based on model • Reweight according to observations • Resample • Particles stay in most probable regions • Performs poorly in hi-dimensional spaces

  15. Incorporating time in First Order Probabilistic Models • Simple approach: Time is one of the arguments in first order logic • Year(p100, 1996), Hot (SVM, 2004) • But time is special • World is growing in the direction of time • Hot (SVM, 2005) dependent on Hot (SVM, 2004) • Hard to discover rules that help in state monitoring, future prediction, etc. • Blowup by incorporating time explicitly • Special inference algorithms no longer applicable

  16. Dynamic Probabilistic Relational Models • DPRM is a PRM replicated over time slices • DBN is a Bayes Net replicated over time slices • In a DPRM attributes for each class dependent on attributes of same/related class • Related class from current/previous time slice • Previous relation • “Unrolled” DPRM = DBN

  17. PLATE1 PLATE1 Color : Red #Holes : 4 Bolted-To : Color : Red #Holes : 4 Bolted-To : BRACKET7 BRACKET7 Color : Blue Size : Large Color : Blue Size : Large Action:Bolt t+1 DPRMs: Example t

  18. Inference in DPRMs • Relational uncertainty  huge state space • E.g. 100 parts  10,000 possible attachments • Particle filter likely to perform poorly • Rao-Blackwellization ?? • Assumptions (relaxed afterwards) • Uncertain reference slots do not appear in slot chains or as parents • Single-valued uncertain reference slots

  19. Rao-Blackwellization in DPRMs • Sample propositional attributes • Smaller space and less error • Constitute the particle • For each uncertain reference slot and particle state • Maintain a multinomial distribution over the set of objects in the target class • Conditioned on values of propositional variables

  20. Pl1 Pl2 Pl3 Pl4 Pl5 Pl1 Pl2 Pl3 Pl4 Pl5 Pl6 Pl7 Pl8 Pl9 Pl10 Blue Small 1lb 0.1 0.1 0.2 0.1 0.5 0.1 0.1 0.2 0.1 0.5 0.4 0.1 0.1 0.3 0.1 Bracket2 RBPF: A Particle Reference slots Propositionalattributes Bolted-To-1 Bolted-To-2 Color Size Wt Pl1 Pl2 Pl3 Pl4 Pl5 Pl1 Pl2 Pl3 Pl4 Pl5 Pl6 Pl7 Pl8 Pl9 Pl10 Red Large 2lbs 0.1 0.1 0.2 0.1 0.5 0.25 0.3 0.1 0.25 0.1 0.3 0.2 0.2 0.1 0.2 Bracket1 ……… ………

  21. Experimental Setup • Assembly Domain (AIPS98) • Objects : Plates, Brackets, Bolts • Attributes : Color, Size, Weight, Hole type, etc. • Relations : Bolted-To, Welded-To • Propositional Actions: Paint, Polish, etc. • Relational Actions: Weld, Bolt • Observations • Fault model • Faults cause uncertainty • Actions and observations uncertain • Governed by global fault probability (fp) • Task: State Monitoring

  22. RBPF vs PF

  23. Problems with DPRMs • Relationships modeled using slots • Slots and slot chains hard to represent and understand • Modeling ternary relationships becomes hard • Small subset of first-order logic (conjunctive expressions) used to specify dependencies • Independence between objects participating in multi-valued slots • Unstructured conditional model

  24. Relational Dynamic Bayes Nets • Replace slots and attributes with predicates (like in MLNs) • Each predicate has parents which are other predicates • The conditional model is a first-order probability tree • The predicate graph is acyclic • A copy of the model at each time slice

  25. Inference: Relaxing the assumptions • RBPF is infeasible when assumptions relaxed • Observation: Similar objects behave similarly • Sample all predicates • Small number of samples, but large relational predicate space • Smoothing : Likelihood of a small number of points can tell relative likelihood of others • Given a particle smooth each relational predicate towards similar states

  26. Simple Smoothing Particle Filtering : A particle Propositionalattributes Bolted-To (Bracket_1, X) Color Size Wt Pl1 Pl2 Pl3 Pl4 Pl5 Pl1 Pl2 Pl3 Pl4 Pl5 Red Large 2lbs 0.1 0.1 0.2 0.1 0.5 1 0 1 1 1 after smoothing Pl1 Pl2 Pl3 Pl4 Pl5 Pl1 Pl2 Pl3 Pl4 Pl5 0.1 0.1 0.2 0.1 0.5 0.9 0.4 0.9 0.9 0.9

  27. Simple Smoothing Problems • Simple smoothing : probability of an object pair depends upon values of all other object pairs of the relation • E.g. P( Bolt(Br_1,Pl_1) ) depends on Bolt(Br_i, Pl_j) for all i and j. • Solution : Make an object pair depend more upon similar pairs • Similarity given by properties of the objects

  28. Abstraction Lattice Smoothing • Abstraction represents a set of similar object pairs. • Bolt(Br1, Pl1) • Bolt(red, large) • Bolt(*,*) • Abstraction Lattice: a hierarchy of abstractions • Each abstraction has a coefficient

  29. Abstraction Lattice : an example

  30. Abstraction Lattice Smoothing • P(Bolt(B1, P1)) = w1 Ppf (Bolt(B1, P1) + w2 Ppf(Bolt(red, large)) + w3 Ppf(Bolt(*,*)) • Joint distributions are estimated using relational kernel density estimation • Kernel K(x, xi) gives distance between the state and the particle • Distance measured using abstractions

  31. Abstraction Smoothing vs PF

  32. Learning with DMLNs • Task: Can MLN learning be directly applied to learn time-based models? • Domains • Predicting author, topic distribution in High-Energy Theoretical Physics papers from KDDCup 2003 • Learning action models of manufacturing assembly processes

  33. Learning with DMLNs • DMLNs = MLNs + Time predicates • R(x,y) -> R(x,y,t), Succ(11, 10), Gt(10,5) • Now directly apply MLN structure learning algorithm (Stanley and Pedro) • To make it work • Use templates to model Markovian assumption • Restrict number of predicates per clause • Add background knowledge

  34. Physics dataset

  35. Manufacturing Assembly

  36. Current and Future Work • Current Work • Programming by Demonstration using Dynamic First Order Probabilistic Models • Future Work • Learning object creation models • Learning in presence of missing data • Modeling hierarchies (very useful for fast inference) • Applying abstraction smoothing to ``static’’ relational models

More Related