1 / 13

Coupling Semi-Supervised Learning of Categories and Relations

Coupling Semi-Supervised Learning of Categories and Relations. by Andrew Carlson, Justin Betteridge , Estevam R. Hruschka Jr. and Tom M. Mitchell School of Computer Science Carnegie Mellon University presented by Thomas Packer. Bootstrapped Information Extraction. Semi-Supervised:

coby
Download Presentation

Coupling Semi-Supervised Learning of Categories and Relations

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Coupling Semi-Supervised Learning of Categories and Relations by Andrew Carlson, Justin Betteridge, EstevamR. HruschkaJr. and Tom M. Mitchell School of Computer Science Carnegie Mellon University presented by Thomas Packer

  2. Bootstrapped Information Extraction • Semi-Supervised: • Seed knowledge (predicate instances & patterns) • Pattern learners (uses learned instances) • Instance learners (uses learned patterns) • Feedback Loop: • Rel1(X, Y) • Sent1(X, Y), Rel0(X, Y)  Pat1 • Pat1: Sent2(A, B)  Rel1(A, B)

  3. Challenges and Previous Solutions • Semantic drift: Feedback loop amplifies error and ambiguities. • Semi-Supervised learning often suffers from being under-constrained. • Multiple mutually-exclusive predicate learning: Positive examples of one predicate are also negative examples of others. • Category and predicate learning: Arguments must be of certain types.

  4. Does More Look Harder?

  5. Approach • Simultaneous bootstrapped training of multiple categories and multiple relations. • Growing related knowledge provides constraints to guide continued learning. • Ontology Constraints: • Mutually exclusive predicates imply negative instances and patterns. • Hypernyms imply positive instances. • Relation argument type constraints imply positive category and negative relation instances.

  6. Mutual Exclusion Constraint • “city” and “scientist” categories are mutually exclusive. • If “Boston” is an instance of “city”, then it is also a negative instance of “scientist”. • If “mayor of arg1” is a pattern for “city”, then it is also a negative pattern for “scientist”.

  7. Hypernym Constraints • “athlete” is a hyponym of “person”. • If “John McEnroe” is a positive instance of athlete, then it is also a positive instance of “person”.

  8. Type Checking Constraints • The “ceoOf()” relation must have arguments of type “person” and “company”. • If “bicycle” is not a “person” then “ceoOf(bicycle, Microsoft)” is a negative instance of “ceoOf()”. • If “ceoOf(Steve Ballmer, Microsoft)” is true, then “Steve Ballmer” is a positive instance of “person”. “Microsoft” handled similarly.

  9. Coupled Bootstrap Learner

  10. Knowledge Constraints Makes Extraction Easier

  11. Knowledge Constraints Makes Extraction Easier

  12. Conclusion • Clearly shows improvements based on constraints. • Could probably benefit by • adding probabilistic reasoning • larger corpus • higher thresholds • more contrastive categories • other techniques discussed in this class

  13. Questions

More Related