1 / 51

Probabilistic Models of Object-Relational Domains

Probabilistic Models of Object-Relational Domains. Daphne Koller Stanford University Joint work with:. Lise Getoor. Ben Taskar. Drago Anguelov. Pieter Abbeel. Rahul Biswas. Nir Friedman. Ming-Fai Wong. Evan Parker. Avi Pfeffer. Eran Segal. Intell_Jane. Diffic_CS101. Intell_George.

sybil
Download Presentation

Probabilistic Models of Object-Relational Domains

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Probabilistic Models of Object-Relational Domains Daphne Koller Stanford University Joint work with: Lise Getoor Ben Taskar Drago Anguelov Pieter Abbeel Rahul Biswas Nir Friedman Ming-Fai Wong Evan Parker Avi Pfeffer Eran Segal

  2. Intell_Jane Diffic_CS101 Intell_George Diffic_CS101 Grade_Jane_CS101 Grade_George_CS101 Intell_George Diffic_Geo101 Grade_George_Geo101 Bayesian Networks: Problem • Bayesian nets use propositional representation • Real world has objects, related to each other Intelligence Difficulty These “instances” are not independent A C Grade

  3. Probabilistic Relational Models • Combine advantages of relational logic & BNs: • Natural domain modeling: objects, properties, relations • Generalization over a variety of situations • Compact, natural probability models • Integrate uncertainty with relational model: • Properties of domain entities can depend on properties of related entities • Uncertainty over relational structure of domain

  4. Teaching-ability Teaching-ability Intelligence Welcome to Geo101 Difficulty Welcome to CS101 Intelligence Difficulty St. Nordaf University Prof. Jones Prof. Smith Teaches Teaches Grade In-course Registered Satisfac George Grade Registered Satisfac In-course Grade Registered Jane Satisfac In-course

  5. Registration Grade Satisfaction Relational Schema • Specifies types of objects in domain, attributes of each type of object & types of relations between objects Classes Student Professor Intelligence Teaching-Ability Teach Take Attributes Relations In Course Difficulty

  6. Representing the Distribution • Very large probability space for a given context • All possible assignments of all attributes of all objects • Infinitely many potential contexts • Each associated with a very different set of worlds Need to represent infinite set of complex distributions

  7. Professor Teaching-Ability Student Intelligence Course Difficulty A B C Reg Grade Satisfaction Probabilistic Relational Models • Universals: Probabilistic patterns hold for all objects in class • Locality: Represent direct probabilistic dependencies • Links define potential interactions [K. & Pfeffer; Poole; Ngo & Haddawy]

  8. Teaching-ability Teaching-ability Grade Intelligence Satisfac Welcome to Geo101 Grade Difficulty Satisfac Welcome to CS101 Intelligence Grade Difficulty Satisfac PRM Semantics • Instantiated PRM BN • variables: attributes of all objects • dependencies: determined by links & PRM Prof. Jones Prof. Smith George Jane

  9. C Welcome to Geo101 Welcome to A low high CS101 The Web of Influence easy / hard low / high

  10. Reasoning with a PRM • Generic approach: • Instantiate PRM to produce ground BN • Use standard BN inference • In most cases, resulting BN is too densely connected to allow exact inference • Use approximate inference: belief propagation • Improvement: Use domain structure — objects & relations — to guide computation • Kikuchi approximation where clusters = objects

  11. Reg Course Student Data  Model  Objects Database Learner Probabilistic Model Expert knowledge • What are the objects in the new situation? • How are they related to each other? Prob. Inference Data for New Situation

  12. Two Recent Instantiations • From a relational dataset with objects & links, classify objects and predict relationships: • Target application: Recognize terrorist networks • Actual application: From webpages to database • From raw sensor data to categorized objects • Laser data acquired by robot • Extract objects, with their static & dynamic properties • Discover classes of similar objects

  13. Summary • PRMs inherit key advantages of probabilistic graphical models: • Coherent probabilistic semantics • Exploit structure of local interactions • Relational models inherently more expressive • “Web of influence”: use multiple sources of information to reach conclusions • Exploit both relational information and power of probabilistic reasoning

  14. Discriminative Probabilistic Models for Relational Data Ben Taskar Stanford University Joint work with: Pieter Abbeel Daphne Koller Ming-Fai Wong

  15. Tom Mitchell Professor Project-of WebKB Project Advisor-of Member Sean Slattery Student Web  KB [Craven et al.]

  16. Student Reg Template potential Intelligence Grade Course Study Group Difficulty Reg2 Student2 Grade Intelligence Undirected PRMs: Relational Markov Nets • Universals: Probabilistic patterns hold for all groups of objects • Locality: Represent local probabilistic dependencies • Address limitations of directed models: • Increase expressive power by removing acyclicity constraint • Improve predictive performance through discriminative training [Taskar, Abbeel, Koller ‘02]

  17. Grade Intelligence Welcome to Difficulty Geo101 Grade Intelligence Grade Difficulty Intelligence Grade RMN Semantics Instantiated RMN  MN • variables: attributes of all objects • dependencies: determined by links & RMN Geo Study Group George Welcome to CS101 CS Study Group Jane Jill

  18. (Reg1.Grade,Reg2.Grade) Grade Grade Intelligence Intelligence Grade Grade Intelligence Intelligence Grade Grade Intelligence Intelligence Grade Grade Learning RMNs • Parameter estimation is not closed form • Convex problem  unique global maximum Maximize L = log P(Grades,Intelligence|Difficulty) easy / hard ABC low / high Grade Difficulty Intelligence Grade Intelligence Grade Difficulty Intelligence Grade

  19. Web Classification Experiments • WebKB dataset • Four CS department websites • Five categories (faculty,student,project,course,other) • Bag of words on each page • Links between pages • Anchor text for links • Experimental setup • Trained on three universities • Tested on fourth • Repeated for all four combinations

  20. To- Page Category ... WordN Word1 Link Exploiting Links From- Page Category ... Classify all pages collectively, maximizing the joint label probability WordN Word1 35.4% relative reduction in error relative to strong flat approach

  21. Scalability • WebKB data set size • 1300 entities • 180K attributes • 5800 links • Network size / school: • 40,000 variables • 44,000 edges • Training time: 20 minutes • Classification time: 15-20 seconds

  22. Member Advisor-of Member Predicting Relationships Tom Mitchell Professor WebKB Project • Even more interesting are relationships between objects Sean Slattery Student

  23. WebKB++ • Four new department web sites: • Berkeley, CMU, MIT, Stanford • Labeled page type (8 types): • faculty, student, research scientist, staff, research group, research project, course, organization • Labeled hyperlinks and virtual links (6 types): • advisor, instructor, TA, member, project-of, NONE • Data set size: • 11K pages • 110K links • 2million words

  24. Flat Model To- Page From- Page ... ... Word1 WordN Word1 WordN Type Rel NONE advisor instructor TA member project-of ... LinkWordN LinkWord1

  25. ... ... ... ... ... ... Flat Model

  26. Collective Classification: Links To- Page From- Page Category Category ... ... Word1 WordN Word1 WordN Type Rel ... LinkWordN LinkWord1

  27. ... ... ... ... ... ... Link Model

  28. Triad Model Advisor Professor Student Member Member Group

  29. Triad Model Advisor Professor Student TA Instructor Course

  30. Triad Model

  31. Error measured over links predicted to be present Link presence cutoff is at precision/recall break-even point (30% for all models) ... ... ... Link Prediction: Results 72.9% relative reduction in error relative to strong flat approach

  32. Summary • Use relational models to recognize entities & relations directly from raw data • Collective classification: • Classifies multiple entities simultaneously • Exploits links & correlations between related entities • Uses web of influence reasoning to reach strong conclusions from weak evidence • Undirected PRMs allow high-accuracy discriminative training & rich graphical patterns

  33. Learning Object Maps from Laser Range Data Dragomir Anguelov Daphne Koller Evan Parker Robotics Lab Stanford University

  34. Occupancy Grid Maps • Static world assumption • Inadequate for answering symbolic queries person robot

  35. Objects • Entities with coherent properties: • Shape • Color • Kinematics (Motion)

  36. Object Maps • Natural and concise representation • Exploit prior knowledge about object models • Walls are straight, doors open and close • Learn global properties of the environment • Primary orientation of walls, typical door width • Generalize properties across objects • Objects viewed as instances of object classes, parameter sharing

  37. Learning Object Maps • Define a probabilistic generative model • Suggest object hypotheses • Optimize the object parameters (EM) • Select highest-scoring model Object Properties Object Segmentation

  38. Laser Sensor Data

  39. Probabilistic Model • Data: A set of scans • Scan: set of <robot position, laser beam reading> tuples • Each scan is associated with a static map M t • Global Map M • A set of objects {1, …, J} • Each object i = {S[i], Dt[i]} • S[i] – static parameters • Dt[i] – dynamic parameters • Non-static environment – dynamic parameters vary only between static maps • Fully dynamic environment – dynamic parameters vary between scans

  40. Probabilistic Model - II • General map M • Static maps M1… MT • Objects i • Robot positions sit • Laser beams zit • Correspondence variables Cit

  41. Generative Model Specification • Sensor model • Object models • Particular instantiation: • Walls • Doors • Model score

  42. Modeling occlusion: Reading ztk generated from: Random model (uniform probability) First object the beam intersects Actual object (Gaussian probability) MaxRange model (Delta function) Why we should model occlusion: Realistic sensor model Helps to infer motion Improved model search Sensor Model

  43. Wall Object Model • Wall model i • A line defined by <i, i>, as in • S intervals <1, 2> each denoting a segment along the line • 2S + 2 independent parameters • Collinear segments bias

  44. Door Object Model • Door Model i • A pivot p • Width w • A set of angles t (t=1,2,…) • Limited rotation (90o) • Arc “center” d • 4 static + 1 dynamic parameter

  45. increased data likelihood Model Score • Define structure prior p(M) over possible maps: • |S[M]| — number of static parameters in M • |D[M]| — number of dynamic parameters in M • L[M] —total length of segments in M Increased number of parameters • Maximize log-posterior probability of map M, data Z:

  46. Learning Model Parameters (EM) • E-step • Compute expectations • M-step • Walls • Optimize line parameters • Optimize segment ends • Doors • Optimize pivot and angles • Optimize door width

  47. Suggesting Object Hypotheses • Wall hypotheses • Use Hough transform (histogram-based approach) • Compute preferred direction of the environment • Use both to suggest lines • Door hypotheses • Use temporal differencing of static maps • Check if points along segments in static maps Mt are well explained in the general map M • If not, the segment is a potential door

  48. Results for a Single Pass

  49. Results for Two Passes

  50. Future Work • Simultaneous localization & mapping • Object class hierarchies • Dynamic environments • Enrich the object representation: • More sophisticated shape models • Color • 3D

More Related