1 / 15

Natural Language Processing

Natural Language Processing. Lecture Notes 12 Chapter 15 (part 2). Harder Example. What makes this hard? What role does Harry play in all this?. Example in class: John told Mary to Walk Fido. Non-Compositionality.

venable
Download Presentation

Natural Language Processing

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Natural Language Processing Lecture Notes 12 Chapter 15 (part 2)

  2. Harder Example • What makes this hard? • What role does Harry play in all this? Example in class: John told Mary to Walk Fido

  3. Non-Compositionality • Unfortunately, there are lots of examples where the meaning (loosely defined) can’t be derived from the meanings of the parts • Idioms, jokes, irony, sarcasm, metaphor, metonymy, indirect requests, etc

  4. English Idioms • Kick the bucket, buy the farm, bite the bullet, run the show, bury the hatchet, etc… • Lots of these… constructions where the meaning of the whole is either • Totally unrelated to the meanings of the parts (kick the bucket) • Related in some opaque way (run the show)

  5. The Tip of the Iceberg • Describe this construction • A fixed phrase with a particular meaning • A syntactically and lexically flexible phrase with a particular meaning • A syntactically and lexically flexible phrase with a partially compositional meaning • …

  6. Example • Enron is the tip of the iceberg. NP -> “the tip of the iceberg” • Not so good… attested examples… • the tip of Mrs. Ford’s iceberg • the tip of a 1000-page iceberg • the merest tip of the iceberg • How about • That’s just the iceberg’s tip.

  7. Example • What we seem to need is something like • NP -> An initial NP with tip as its head followed by a subsequent PP with of as its head and that has iceberg as the head of its NP And that allows modifiers like merest, Mrs. Ford, and 1000-page to modify the relevant semantic forms

  8. Constructional Approach • Syntax and semantics aren’t separable in the way that we’ve been assuming • Grammars contain form-meaning pairings that vary in the degree to which the meaning of a constituent can be computed from the meanings of the parts.

  9. Constructional Approach • So we’ll allow both VP →V NP {V.sem(NP.sem)} and VP →Kick-Verb the bucket {λx Die(x)}

  10. Computational Realizations • Semantic grammars • Simple idea, misleading name • Read for your interest: information extraction, cascaded finite-state transducers

  11. Semantic Grammars • Traditional grammars may not reflect the semantics in a straightforward way • You can deal with this by… • Fighting with the grammar • Complex lambdas and complex terms, etc • Rewriting the grammar to reflect the semantics • And in the process give up on some syntactic niceties

  12. BERP Example

  13. BERP Example • How about a rule like the following instead… Request → I want to go to eat FoodType Time { some attachment }

  14. Semantic Grammar • The technology (plain CFG rules with a set of terminals) is the same as we’ve been using • Good: you get exactly the semantic rules you need • Bad: you need to develop a new grammar for each new domain

  15. Semantic Grammars • Typically used in conversational agents in constrained domains • Limited vocabulary • Limited grammatical complexity • Chart parsing (Earley) can often produce all that’s needed for semantic interpretation even in the face of ungrammatical input.

More Related