1 / 31

Information Extraction

Information Extraction. Feiyu Xu Jakub Piskorski DFKI LT-Lab. Overview. Introduction to information extraction SPPC Survey of information extraction approaches MUC conferences. What is Information Extraction.

watson
Download Presentation

Information Extraction

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Information Extraction Feiyu Xu Jakub Piskorski DFKI LT-Lab

  2. Overview • Introduction to information extraction • SPPC • Survey of information extraction approaches • MUC conferences

  3. What is Information Extraction The creation of a structured representation (TEMPLATE) of selected information drawn from natural language text Task is more limited than full text understanding: extracting relevant information, ignoring irrelevant information

  4. Example(Management Succession) • Dr. Hermann Wirth, bisheriger Leiter der Musikhochschule München, verabschiedete sich heute aus dem Amt. • Der 65jährigetritt seinen wohlverdienten Ruhestand an. • Als seine NachfolgerinwurdeSabine Klingerbenannt. • Ebenfalls neu besetzt wurde die Stelle des Musikdirektors. • Annelie Häfnerfolgt Christian Meindlnach. PersonOut: Dr. Hermann Wirth PersonIn: Sabine Klinger Position: Leiter Organization: Musikhochschule München TimeOut: Heute TimeIn: PersonOut: Christian Meindl PersonIn: Annelie Häfner Position: Musikdirektor Organization: ? TimeOut: TimeIn:

  5. Template filling rule VG:verabschiedet sich aus dem Amt Subject:[1] NP NLP: PERSON: [2] Domain: IN: ORGANISATION: [3] POSITION: [4] OUT: [1] PERSON: ORGANISATION: [3] POSITION: [4]

  6. Template filling rule VG:verabschiedete sich heute aus dem Amt Subject:[1] NLP: PERSON_NAME: [5] POSITION_NAME: [4] ORGANISATION_NAME: [3] PERSON: [2] Domain: IN: ORGANISATION: [3] POSITION: [4] [1] PERSON: OUT: ORGANISATION: [3] POSITION: [4]

  7. Tokenization Traditional IE Architecture Text Sectionizing and Filtering Morphological and Lexical Processing Local Text Analysis Part of Speech Tagging Word Sense Tagging Fragment Processing Fragment Combination Scenario Pattern Matching Parsing Discourse Analysis Coreference Resolution Inference Template Merging

  8. Introduction SPPC - SHALLOW PROCESSING PRODUCTION CENTER • System for shallow analysis of German free text documents (toolset of integrated shallow components) • based on shallow text processor of SMES (G. Neumann) • exhaustive use of finite-state technology: - DFKI Finite-State Machine Toolkit - dynamic tries • high linguistic coverage • very good performance

  9. XML Documents ASCII Documents Tokenizer Lexical Processor POS-Filtering Named Entity Finder Phrase Recognizer Sentence Boundary Detection Architecture EXAMPLE: rund 60 bis 70 Prozent der Steigerungsrate (about 60 to 70 percent increase) rund: lowercase 60: two-digit-integer Steigerungsrate: steigerung+[s]+rate bis: prep|adv Linguistic Knowledge Pool bis: adv rund 60 bis 70 Prozent: percentage-NP Text Chart rund 60 bis 70 Prozent: NP der Steigerungsrate: NP XML-Output Interface

  10. Tokenizer • The goal of the TOKENIZER is to: • - map sequences of consecutive characters into word-like units (tokens) - identify the type of each token disregarding the context - performing word segmentation when necessary (e.g., splitting contractions into multiple tokens if necessary) • overall more than 50 classes (proved to simplify processing on higher stages) - NUMBER_WORD_COMPOUND (“69er”) - ABBREVIATION (CANDIDATE_FOR_ABBREVIATION), - COMPLEX_COMPOUND_FIRST_CAPITAL („AT&T-Chief“) - COMPLEX_COMPOUND_FIRST_LOWER_DASH („d‘Italia-Chefs-“) • represented as single WFSA (406 KB)

  11. Lexical Processor • Tasks of the LEXICAL PROCESSOR: • - retrieval of lexical information- recognition of compounds („Autoradiozubehör“ - car-radio equipment)- hyphen coordination („Leder-, Glas-, Holz- und Kunstoffbranche“ leather, glass, wooden and synthetic materials industry) • lexicon contains currently more than 700 000 German full-form words (tries) • each reading represented as triple <STEM,INFLECTION,POS> • example: „wagen“ (to dare vs. a car) STEM: „wag“INFL: (FORM: infin) (TENSE: pres, PERSON: anrede, NUMBER: sg) (TENSE: pres, PERSON: anrede, NUMBER: pl) (TENSE: pres, PERSON: 1, NUMBER: pl) (TENSE: pres, PERSON: 3, NUMBER: pl) (TENSE: subjunct-1, PERSON: anrede, NUMBER: sg) (TENSE: subjunct-1, PERSON: anrede, NUMBER: pl) (TENSE: subjunct-1, PERSON: 1, NUMBER: pl) (TENSE: subjunct-1, PERSON: 3, NUMBER: pl) (FORM: imp, PERSON: anrede) POS: noun STEM: „wag“INFL: (GENDER: m,CASE: nom, NUMBER: sg) (GENDER: m,CASE: akk, NUMBER: sg) (GENDER: m,CASE: dat, NUMBER: sg) (GENDER: m,CASE: nom, NUMBER: pl) (GENDER: m,CASE: akk, NUMBER: pl) (GENDER: m,CASE: dat, NUMBER: pl) (GENDER: m,CASE: gen, NUMBER: pl)POS: noun

  12. Part-of-Speech Filtering • The task of POS FILTER is to filter out unplausible readings of ambiguous word forms • large amount of German word forms are ambiguous (20% in test corpus) • contextual filtering rules (ca. 100) • - example: „Sie bekannten, die bekannten Bilder gestohlen zu haben“They confessed they have stolen the famous pictures „bekannten“ - to confess vs. famous • FILTERING RULE: if the previous word form is determiner and the nextword form is a noun then filter out the verb reading of the current word form • supplementary rules determined by Brill‘s tagger in order to achieve broader coverage • rules represented as FSTs, hard-coded rules (filtering out rare readings)

  13. Named Entity Recognition • The task of NAMED ENTITY FINDER is the identification of: • - entities: organizations, persons, locations - temporal expressions: time, date - quantities: monetary values, percentages, numbers • identification of named entities in two steps: • - recognition patterns expressed as WFSA are used to identify phrases containing potential candidates for named entities (longest match strategy) • - additional constraints (depending on the type of candidate) are used for validating the candidates • on-line base lexicon for geographical names, first names(persons)

  14. Named Entity Recognition • since named entities may appear without designators (companies, persons) adynamic lexicon for storing such named entities is used • candidates for named entities: • example: • Da flüchten sich die einen ins Ausland, wie etwa der Münchner Strickwarenhersteller März GmbH oder der badische Strumpffabrikant Arlington Socks, GmbH. Ab kommendem Jahr strickt März knapp drei Viertel seiner Produktion in Ungarn. • partial reference resolution (acronyms) • resolution of type ambiguity using the dynamic lexicon: • example: • if an expression can be a person name or company name (Martin Marietta Corp.) then use type of last entry inserted into dynamic lexicon for making decision

  15. Phrase Recognition • The task of FRAGMENT RECOGNIZER is to recognize nominal and prepositional phrases and verb groups • example: [NPDas Unternehmen][ORG Merck]][V geht] [DATE im Herbst 1999] [PP mit einem Viertel seines Kapitals] [PP an die[ORG Frankfurter Börse]] „In autumn 1999 the company Merck will go public to the Frankfurt Stock Exchange with one fourth of it‘s capital “ • extraction patterns expressed as WFSAs (mainly based on POS information and named entity type) • phrasal grammars only consider continuous substrings (recognition of verb groups is partial) • example: „Gestern [ist] der Linguist vom neuen Manager [entlassen worden].“Yesterday, the linguist [has][been fired] by the new manager. Sentence Boundary Detection • few simple contextual rules are sufficient since named entity recognition is performed earlier

  16. Evaluation & Performance Performance: INPUT: corpus of German business magazine „Wirtschaftswoche“ (1,2MB) TIME: ~15sec. ( ~13400 wrds/sec) on Pentium III, 850MHz, 256 RAM RECOGNIZED: 197118 tokens, 11575 named entities, 64839 Phrases Evaluation: Recall Precision COMPOUND ANALYSIS: 98.53% 99.29% POS-FILTERING: 74.50% 96.36% NE-RECOGNITION: 85% 95.77% FRAGMENTS (NPs,PPs): 76.11% 91.94% Availability: C++ Library for UNIX, Windows 98/NT, Demo with GUI for Windows 98/NT

  17. Two Approaches to Building IE Systems[Appelt & Israel, 99] • Knowledge Engineering Approach • Grammars are constructed by hand • Domain patterns are discovered by a human expert through introspection and inspection of a corpus • Much laborious tuning • Automatically Trainable Systems • Use statistical methods when possible • Learn rules from annotated corpora, e.g., • statistical name recognizer • Learn rules from interaction with user, e.g., • learning template filler rules

  18. Knowledge Engineering[Appelt & Israel, 99] • Advantages • With skill and experience, good performing systems are not conceptually hard to develop • The best performing systems have been hand crafted • Disadvantages • Very laborious development process • Some changes to specifications can be hard to accommodate, e.g., • Name recognition rules based upper and lower cases • Required experts may not be available

  19. Trainable Systems[Appelt & Israel, 99] • Advantages • Domain portability is relatively straightforward • System experts is not required for customization • “Data driven” rule acquisition ensures full coverage of examples • Disadvantages • Training data may not exist, and may be very expensive to acquire • Large volume of training data may be required • Changes to specifications may require reannotation of large quantities of training data

  20. When Works Best for Knowledge based IE?[Appelt & Israel, 99] • Resources (e.g. lexicons, lists) are available • Rule writers are available • Training data scarce or expensive to obtain • A mixture of knowledge based and machine learning based approach is also possible!

  21. “pseudo-syntax” [Hobbs et al. 96] • The material between the end of the subject noun group and the beginning of the main verb group must be read over, for example, • Read over prepositional phrases and relative clauses • Subject {Preposition NounGroup}* VerbGroup • Subject Relpro {NounGroup | Other }* VerbGroup The mayor, who was kidnapped yesterday, was found dead today. • Conjoined verb phrase, skipping over the first conjunct and associate the subject with the verb group in the second conjunct Subject VerbGroup {NounGroup|Other}* Conjunction VerbGroup

  22. Same semantic content can be realized in different forms. GM manufactures cars. Cars are manufactured by GM. ... GM, which manufactures cars ... ... cars, which are manufactured by GM ... ... cars manufactured by GM ... GM is to manufacture cars. Cars are to be manufactured by GM. GM is a car manufacturer. • Question: How many rules are needed to extract all relevant patterns? Why not using a linguistic theory? • Performance vs. Compotence • TACITUS: 36 hours to process 100 Messages • FASTUS: 12 minutes to process 100 messages Problem of “pseudo-syntax” [Hobbs et al. 96]

  23. Message Understanding Conferences[MUC-7 98] • U.S. Government sponsored conferences with the intention to coordinate multiple research groups seeking to improve IE and IR technologies (since 1987) • defined several generic types of information extraction tasks(MUC Competition) • MUC 1-2 focused on automated analysis of military messages containing textual information • MUC 3-7 focused on information extraction from newswire articles • terrorist events • international joint-ventures • management succession event

  24. Evaluation of IE systems in MUC • Participants receive description of the scenario along with the annotated training corpus in order to adapt their systems to the new scenario (1 to 6 months) • Participants receive new set of documents (test corpus) and use their systems to extract information from these documents and return the results to the conference organizer • The results are compared to the manually filled set of templates (answer key)

  25. Evaluation of IE systems in MUC • precision and recall measures were adopted from the information retrieval research community • Sometimes an F-meassure is used as a combined recall-precision score

  26. Generic IE tasks for MUC-7 • (NE) Named Entity Recognition Task requires the identification an classification of named entities • organizations • locations • persons • dates, times, percentages and monetary expressions • (TE) Template Element Task requires the filling of small scale templates for specified classes of entities in the texts • Attributes of entities are slot fills (identifying the entities beyond the name level) • Example: Persons with slots such as name (plus name variants), title, nationality, description as supplied in the text, and subtype.“Capitan Denis Gillespie, the comander of Carrier Air Wing 11”

  27. Generic IE tasks for MUC-7 • (TR) Template Relation Task requires filling a two slot template representing a binary relation with pointers to template elements standing in the relation, which were previously identified in the TE task • subsidiary relationship between two companies(employee_of, product_of, location_of)

  28. Generic IE tasks for MUC-7 • (CO) Coreference Resolution requires the identification of expressions in the text that refer to the same object, set or activity • variant forms of name expressions • definite noun phrases and their antecedents • pronouns and their antecedents • “The U.K. satellite television broadcaster said its subscriber base grew 17.5 percentduring the past year to 5.35 million” • bridge between NE task and TE task

  29. Generic IE tasks for MUC-7 • (ST) Scenario Template requires filling a template structure with extracted information involving several relations or events of interest • intended to be the MUC approximation to a real-world information extraction problem • identification of partners, products, profits and capitalization of joint ventures

  30. Tasks evaluated in MUC 3-7[Chinchor, 98]

  31. Maximum Results Reported in MUC-7

More Related