1 / 27

QuASI: Question Answering using Statistics, Semantics, and Inference

QuASI: Question Answering using Statistics, Semantics, and Inference. Marti Hearst, Jerry Feldman, Chris Manning, Srini Narayanan Univ. of California-Berkeley / ICSI / Stanford University. Outline. Project Overview Three topics: Assigning semantic relations via lexical hierarchies

albany
Download Presentation

QuASI: Question Answering using Statistics, Semantics, and Inference

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. QuASI:Question Answering using Statistics, Semantics, and Inference Marti Hearst, Jerry Feldman, Chris Manning, Srini Narayanan Univ. of California-Berkeley / ICSI / Stanford University

  2. Outline • Project Overview • Three topics: • Assigning semantic relations via lexical hierarchies • From sentences to meanings via syntax • From text analysis to inference using conceptual schemas

  3. Main Goals Support Question-Answering and NLP in general by: • Deepening our understanding of concepts that underlie all languages • Creating empirical approaches to identifying semantic relations from free text • Developing probabilistic inferencing algorithms

  4. Two Main Thrusts • Text-based: • Use empirical corpus-based techniques to extract simple semantic relations • Combine these relations to perform simple inferences • “statistical semantic grammar” • Concept-based: • Determine language-universal conceptual principles • Determine how inferences are made among these

  5. Relation Recognition Abbreviation Definition Recognition Semantic Relation Identification

  6. UCB, Sept-Nov, 2002 • Abbreviation Definition Recognition • Developed and evaluated new algorithm • Better results than existing approaches • Simpler and faster as well • Semantic Relation Identification • Developed syntactic chunker • Analyzed sample relations • Began development of a new computational model • Incorporates syntax and semantic labels • Test example: identify “treatment for disease”

  7. Abbreviation Examples • “Heat-shock protein 40 (Hsp40) enables Hsp70 to play critical roles in a number of cellular processes, such as protein folding, assembly, degradation and translocation in vivo.” • “Glutathione S-transferase pull-down experiments showed the direct interaction of in vitro translated p110, p64, and p58 of the essential CBF3 kinetochore protein complex with Cbf1p, a basic region helix-loop-helix zipper protein (bHLHzip) that specifically binds to the CDEI region on the centromere DNA.” • “Hpa2 is a member of the Gcn5-related N-acetyltransferase (GNAT) superfamily, a family of enzymes with diverse substrates including histones, otherproteins,arylalkylamines and aminoglycosides.”

  8. Related Work • Pustejovsky et al. present a solution based on hand-build regular expression and syntactic information. Achieved 72% recall at 98% • Chang et al. use linear regression on a pre-selected set of features. Achieved 83% recall at 80%* precision, and 75% recall at 95% precision. • Park and Byrd present a rule-based algorithm for extraction of abbreviation definitions in general text. • Yoshida et al. present an approach close to ours, trying to first match characters on word and syllable boundaries. * Counting partial matches, and abbreviations missing from the “gold-standard” their algorithm achieved 83% recall at 98% precision.

  9. The Algorithm • Much simpler than other approaches. • Extracts abbreviation-definition candidates adjacent to parentheses. • Finds correct definitions by matching characters in the abbreviation to characters in the definition, starting from the right. • The first character in the abbreviation must match a character at the beginning of a word in the definition. • To increase precision a few simple heuristics are applied to eliminate incorrect pairs. • Example: Heat shock transcription factor (HSF). • The algorithm finds the correct definition, but not the correct alignment: Heat shock transcription factor

  10. Results • On the “gold-standard” the algorithm achieved 83% recall at 96% precision.* • On a larger test collection the results were 82% recall at 95% precision. • These results show that a very simple algorithm produces results that are comparable to these of the exiting more complex algorithms. * Counting partial matches, and abbreviations missing from the “gold-standard” our algorithm achieved 83% recall at 99% precision.

  11. From sentences to meanings via syntax Factored A* Parsing Relational approaches to semantic relations Learning aspectual distinctions

  12. Factored A* Parsing • Goal: develop a lexicalized parser that is fast, accurate and exact[finds the model’s best parse] • Technology exists to get any two, but not all three • Approximate Parsing – Fast but Inexact • Beam or “Best-First” Parsing [Charniak, Collins, etc.] • Factored: represent tree and dependencies separately • Simple, modular, extensible design • Permits fast, high accuracy, exact inference • A* estimates combined from product of experts model • Available from: http://nlp.stanford.edu/[Java, src]

  13. T L D Factored A* Parsing Dependency Accuracy Work Done Labeled Bracketing Accuracy (F1)

  14. Learning Semantic Relations • FrameNet as starting point and training data • Constraint Resolution for Entire Relations • Logical Relations • Probabilistic Models • Combinations of the Two • Bootstrap to new domains • Building blocks for Q/A relevant tasks: • Semantic Roles in Text • Inference • Improved Syntactic Parsing

  15. John has lived in Miami for ten years now. Event Ref-time Event John has lived in Miami before. Learning Aspect:The Perfect • English perfect has experiential, relevant, and durative readings: have been to Bali vs. have just eaten lunch • Disambiguation is necessary for text understanding: John has traveled to Malta [now, or in the past?] • Siegel (2000) looked at inherent but not contextual aspect • Current status: annotation underway for training statistical classifier

  16. Concept-based Analysis From text analysis to inference using conceptual schemas Relational Probabilistic Models Open Domain Conceptual Relations

  17. Hypothesis: Linguistic input is converted into a mental simulation based on bodily-grounded structures. Components: Semantic schemas image schemas and executing schemas are abstractions over neurally grounded perceptual and motor representations Linguistic units lexical and phrasal construction representations invoke schemas, in part through metaphor Inferencelinks these structures and provides parameters for a simulation engine Inference and Conceptual Schemas

  18. Conceptual Schemas • Much is known about conceptual schemas, particularly image schemas • However, this understanding has not yet been formalized • We will develop such a formalism • They have also not been checked extensively against other languages • We will examine Chinese, Russian, and other languages in addition to English

  19. Schema Formalism SCHEMA <name> SUBCASE OF <schema> EVOKES <schema> AS <local name> ROLES < self role name>: <role restriction> < self role name> <-> <role name> CONSTRAINTS <role name> <- <value> <role name> <-> <role name> <setting name> :: <role name> <-> <role name> <setting name> :: <predicate> | <predicate>

  20. A Simple Example SCHEMA hypotenuse SUBCASE OF line-segment EVOKES right-triangle AS rt ROLES Comment inherited from line-segment CONSTRAINTS SELF <-> rt.long-side

  21. Source-Path-Goal SCHEMA: spg ROLES: source: Place path: Directed Curve goal: Place trajector: Entity

  22. Extending Inferential Capabilities • Given the formalization of the conceptual schemas • How to use them for inferencing? • Earlier pilot systems • Used metaphor and Bayesian belief networks • Successfully construed certain inferences • But don’t scale • New approach • Probabilistic relational models • Support an open ontology

  23. A Common Representation • Representation should support • Uncertainty, probability • Conflicts, contradictions • Current plan • Probabilistic Relational Models (Koller et al.) • DAML + OIL

  24. Status of PRM for AQUAINT • Fall 2002 • Designed the basic PRM code-base/infrastructure • Packages for BN’s, OOBN. • Designed PRM inference Algorithm. • Spring-Summer 2003 • Implement the PRM inference Algorithm • Design Dynamic Probabilistic Relational Models (DPRM) • Implement DPRM to replace Pilot System DBN • Test DPRM for QA • Related Work • Probabilistic OWL (PrOWL) • Probabilistic FrameNet

  25. An Open Ontology for Conceptual Relations • Build a formal markup language for conceptual schemas • We propose to use DAML+OIL/OWL as the base. • Advantages of the approach • Common framework for extending and reuse • Closer ties to other efforts within AQUAINT as well as the larger research community on the Semantic Web. • Some Issues • Expressiveness of DAML+OIL • Representing Probabilistic Information • Extension to MetaNet, capture abstract concepts

  26. Current Status • Summer/Fall 2002 • FrameNet-1 is available in DAML+OIL • http://www.icsi.berkeley.edu/~framenet • Image Schemas have been formalized and DAML+OIL representation designed • Initial set of Metaphors and an SQL Metaphor database is in place. • Spring 2003 • Populate Metaphor Database • Populate Image Schema Database • Summer 2003 • Test Inferencing with Image Schemas for QA.

  27. Putting it all Together • We have proposed two different types of semantics • Universal conceptual schemas • Semantic relations • In Phase I they will remain separate • However, we are exploring using PRMs as a common representational format • In later Phases they will be combined

More Related