1 / 28

Relation Extraction

Relation Extraction. Pierre Bourreau LSI-UPC PLN-PTM. Plan . Relation Extraction description Sampling templates Reducing deep analysis errors… Conclusion. Relation Extraction Description. Finding relations between entities into a text Filling pre-defined templates slots

tender
Download Presentation

Relation Extraction

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Relation Extraction Pierre Bourreau LSI-UPC PLN-PTM

  2. Plan • Relation Extraction description • Sampling templates • Reducing deep analysis errors… • Conclusion

  3. Relation Extraction Description • Finding relations between entities into a text • Filling pre-defined templates slots • One-value-per-field • Multi-value • Depend on analysis: • Chunking • Tokenization • Sentence Parsing…

  4. Plan • Relation Extraction description • Sampling templates (Cox, Nicolson, Finkel, Manning) • Reducing deep analysis errors… • Conclusion

  5. First Example: Sampling Templates • Example: workshop announcement • PASCAL corpus • Relations to extract: • dates of events • Workshop conferences names, acronyms and URL • Domain knowledge: • Constraints on dates • Constraints on names

  6. PASCAL Corpus: semi-structured corpus • <0.26.4.95.11.09.31.hf08+@andrew.cmu.edu.0> • Type: cmu.andrew.academic.bio • Topic: "MHC Class II: A Target for Specific Immunomodulation of the • Immune Response" • Dates: 3-May-95 • Time: <stime>3:30 PM</stime> • Place: <location>Mellon Institute Conference Room</location> • PostedBy: Helena R. Frey on 26-Apr-95 at 11:09 from andrew.cmu.edu • Abstract: • Seminar: Departments of Biological Sciences • Carnegie Mellon and University of Pittsburgh • Name: <speaker>Dr. Jeffrey D. Hermes</speaker> • Affiliation: Department of Autoimmune Diseases Research & Biophysical Chemistry • Merck Research Laboratories • Title: "MHC Class II: A Target for Specific Immunomodulation of the • Immune Response" • Host/e-mail: Robert Murphy, murphy@a.cfr.cmu.edu • Date: Wednesday, May 3, 1995 • Time: <stime>3:30 p.m.</stime> • Place: <location>Mellon Institute Conference Room</location> • Sponsor: MERCK RESEARCH LABORATORIES • Schedule for 1995 follows: (as of 4/26/95) • Biological Sciences Seminars 1994-1995 • Date Speaker Host • April 26 Helen Salz Javier L~pez • May 3 Jefferey Hermes Bob Murphy • MERCK RESEARCH LABORATORIES

  7. PASCAL Corpus: semi-structured corpus • <1.21.10.93.17.00.39.rf1u+@andrew.cmu.edu.0> • Type: cmu.andrew.org.heinz.great-lake • Topic: Re: PresentationCC: • Dates: 25-Oct-93 • Time: <stime>12:30</stime> • PostedBy: Richard Florida on 21-Oct-93 at 17:00 from andrew.cmu.edu • Abstract: • Folks: • <paragraph> <sentence>Our client has requested that the presentation be postponed until Monday • during regular class-time</sentence>. <sentence>He has been asked to make a presentaion for • the Governor of Michigan and Premier of Ontario tommorrow morning in • Canada, and was afraid he could not catch a plane in time to make our • presentation</sentence>. <sentence>After consulting with Rafael and a sub group of project • managers, it was decided that Monday was the best feasible presentation • alternative</sentence>. <sentence>Greg has been able to secure Room 2503 in Hamburg Hall for • our presentation Monday during regular class-time</sentence>. </paragraph> • <paragraph><sentence>We will meet tommmorow in <location>2110</location> at <stime>12:30</stime> (lunch provided) to finalize • presentation and briefing book</sentence>. <sentence>Also, the client has faxed a list of • reactions and questions for discussion which we should review</sentence>. • <sentence>Thanks very much for your hard work and understanding</sentence>. <sentence>Look forward to • seeing you tommorrow</sentence>.</paragraph> • Richard

  8. Idea • Sampling Templates: • Generate all available templates • Give a probability to each of them • Relational model: • Constraints on dates: order • 1. submission dates • 2. acceptance dates • 3. workshop dates / camera ready dates • Constraints on names. • Slots: name, acronym, URL • URL is generated from acronyms

  9. Baselines • CRF • Cliques: max=2 • Viterbi algorithm • Token => GATE tokenization • CMM • Idem • Window of the four previous tokens

  10. Templates sampling • Tokens  p(Li|Li-1) or p(Li|Li-1,…, Li-4) on 100 of documents • Template: • Each slot holds one/no filler value • -> date templates: • SUB_DATE • ACC_DATE • WORK_DATE • CAMREADY_DATE

  11. Templates sampling • Tokens  p(Li|Li-1) or p(Li|Li-1,…, Li-4) on 100 of documents • Template: • Each slot holds one/no filler value • -> name templates: • CONF_NAME • CONF_ACRO • CONF_URL • WORK_NAME • WORK_ACRO • WORK_URL

  12. Templates sampling • D a distribution of these templates, over the training set. => LOCAL MODEL (PL)

  13. Templates scoring: Date Model • PA/P: Probability of present/absent fields. Set with training data • Po: Ordering probability. We give penalty to constraints violations. • PA/P* Po = Prel

  14. Templates scoring: Name Model • Name->Acronym: independent module (likelihood score – Chang 2002): Pnam->acr • Acronym->URL: empirical probability from training: Pacr->url • Pb: missing entry give advantage to incomplete templates. • PA/P: pondering templates (in training, most values are filled) • Prel= Pnam->acr *Pacr->url *PA/P

  15. Results: 300 documents

  16. Results • No results over CRF • CRF accepts variation (ex: name) • => lower recall • Rel. Model does not improve CRF (not on graph) • Low-window of CRF => less info in distribution. • Substantial improvement over CMM (5%)

  17. Plan • Relation Extraction description • Sampling templates • Reducing deep analysis errors (Zao, Grishman) • Conclusion

  18. Problematic • Use different syntactic analysis for the task: • Tokenization • Chunking • Sentence Parsing • … • The more info they give, the less accurate they are. • =>combine them to correct errors

  19. ACE task… remember • Entities: • PERson – ORGanisation – FACility – GeoPoliticEntity - LOCation – WEApon – VEHicle • Mentions: • NAM (proper), NOM (nominal), PRO (pronoun) • Relations: • EMP-ORG, PHYS, GPE-AFF, PER-SOC, DISC, ART, Other

  20. Kernel, SVM … nice properties • Kernel: • Function replacing scalar vector products • Enables us to translate problems into a higher-dimension space for solution • Sum, product generates kernels. • SVM: • SVM can pick up features for best separation

  21. The relational model • R=(arg1, arg2, seq, link, path) • arg1, arg2: the two entities to compare • seq=(t1, …, tn): sequence of tokens intervening • link=(t1, …, tm): idem seq but just with important words • path: a dependency path… • T=(word, pos, base) • pos: Part Of Speech tagging • base: morphological base • E=(tk, type, subtype, mtype) • type: according to ACE type • subtype: refining • mtype: the way it is mentioned • DT=(T, dseq) • dseq=(arc1, …, arcn) • ARC=(w, dw, label, e) • w: current token • dw: token connected to w • label: role label of this arc • e: direction of the arc

  22. The relational model: example arg1=((“areas”, “NNS”, “area”, dseq), “LOC”, “region”, “NOM”) • arg1.dseq=((OBJ, areas, in, 1), (OBJ, areas, controlled, 1)) path=((OBJ, areas, controlled, 1), (SBJ, controlled, troops, 0))

  23. Kernels • Argument kernel: • Matches two tokens, comparing each fix arguments (word, pos, type…) • Bigram kernel: • Matches token on a window of size 1 • Link sequence kernel: • Relations often occur in a short context.

  24. Kernels (2) 4. Dependency path kernel: • How similar are two paths? 5. Local dependency kernel: • Idem as path but more informative. • Helpful if dependency path does not exist.

  25. Results: adding info into SVM • The more information we give, the better the result. • Link Sequence Kernel boosts results.

  26. Results: SVM or KNN • SVM behaves globally better • Polynomial extension has no consequence on KNN. • Training problem in the last three. • … good results over ACE official task… secret, no comparison available

  27. Conclusion • Really simple method • Nice properties of Kernel/SVMs • This method is generic!!! (tested on annotated text) • Looks like SVM can process better, for this task. • … but hard to compare the two methods as goals are different.

  28. References • [1] Template Sampling for Leveraging Domain Knowledge in Information Extraction. Cox, Nicolson, Finkel, Manning, Langley. Stanford University. • [2] Extracting Relations with Integrated Information Using Kernel Methods. Zao, Grishman. New York University. 2005

More Related