1 / 74

Semantic Role Labeling of Implicit Arguments for Nominal Predicates

Language & Interaction Research. Semantic Role Labeling of Implicit Arguments for Nominal Predicates. Matthew Steven Gerber. Department of Computer Science Michigan State University East Lansing, Michigan, USA. The Semantic Role. The relationship between a noun phrase and a predicate

clio
Download Presentation

Semantic Role Labeling of Implicit Arguments for Nominal Predicates

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Language & Interaction Research Semantic Role Labelingof Implicit Arguments forNominal Predicates Matthew Steven Gerber Department of Computer Science Michigan State University East Lansing, Michigan, USA

  2. The Semantic Role • The relationship between a noun phrase and a predicate • John threw a ball to Mary. • A ball was thrown to Mary by John. • … • John: the thrower • a ball: entity thrown • Mary: recipient • Semantically normalize syntactic variation semantic roles

  3. Semantic Role Labeling (SRL) • Automatically identify semantic roles • Example • John shipped a package from Michigan to California. • Benefits downstream processes • Question answering (Pizzato and Molla, 2008) • Information extraction (Banko and Etzioni, 2008) • Machine translation (Liu and Gildea, 2010) [Sender John] [Predicate shipped] [Thingshippeda package] [Source from Michigan] [Destination to California].

  4. Semantic Role Labeling (SRL) • 2002 – present: mostly focused on verbal SRL • Noun-based (nominal) SRL is also possible • Example • John made a shipment from Michigan to California. [Sender John] made a [Predicate shipment] [Source from Michigan] [Destination to California].

  5. Arg0: distributor • Arg1: thing distributed • Arg2: recipient Nominal SRL Based on NomBank • Lexicon entry • Distribution: send a package • Annotation • Searle will give pharmacists brochures on the use of prescription drugs for distribution in their stores. • Automatic nominal SRL (Liu and Ng, 2007) • Supervised machine learning • Test over predicates known to take arguments Searle will give [Arg0 pharmacists] [Arg1 brochures on the use of prescription drugs] for [Predicate distribution] [Location in their stores].

  6. Problems due to Implicit Arguments John threw a ball to Mary. The catch was amazing. • Evaluation methodology • Test predicates with local arguments • Not all predicates have local arguments • Does not reflect practical system performance • Implicit arguments • Not annotated by NomBank • Often exist in surrounding text • Can they be automatically recovered?

  7. Research Contributions Investigate implicit arguments (IAs) in nominal SRL • Examine the effect of IAs on nominal SRL performance • Improve SRL performance by taking IAs into account • Recover implicit arguments from discourse

  8. Research Contributions Investigate implicit arguments (IAs) in nominal SRL • Examine the effect of IAs on nominal SRL performance • Improve SRL performance by taking IAs into account • Recover implicit arguments from discourse

  9. Nominal SRL Model • Full syntactic analysis S VP S VP VP NP NP PP NP JJ NNS Judge Curry ordered Edison to make average [Predicate refunds] of about $45.

  10. Nominal SRL Model • Full syntactic analysis • 23-class classification problem over parse tree nodes S VP S VP VP NP NP PP NP JJ NNS Judge Curry ordered [Arg0 Edison] to make average [Predicate refunds] [Arg1 of about $45].

  11. Nominal SRL Model • Features • Primarily lexical and syntactic • Most important feature: parse tree path

  12. NNS VP S VP … Nominal SRL Model • NP S VP S VP VP NP NP PP NP JJ NNS Judge Curry ordered [ ? Edison] to make average [Predicate refunds] of about $45.

  13. NNS:refund Nominal SRL Model • Other notable features • Parse path concatenated with predicate • First word of candidate argument • Greedy forward feature selection • 31 selected features • 2 numeric features • Other features are binarized VP S VP … • NP

  14. Evaluation of Nominal SRL Model • Training data • Sections 2-21 of NomBank • 3.7 million training instances • Development data • Section 24 • Feature selection • Threshold tuning • Testing data • Section 23

  15. Evaluation of Nominal SRL Model • Results • Previous best argument F1: 72.8% (Liu and Ng, 2007) • Our best argument F1: 75.7% (Gerber et al., 2009) • Key differences • New features (e.g., lexicalized parse path) • Single-stage classification pipeline • Problem: ground-truth predicates • In a realistic setting, we won’t have this information

  16. Impact of Implicit Arguments • Extended evaluation • Process each token in test data • Attempt SRL for known nouns • Results

  17. Impact of Implicit Arguments • Example error • Canadian investment rules require that big foreign takeovers meet that standard. • Compare to • Canadian investment in the US has declined. [Arg0 Canadian] [Predicate investment] rules require that big foreign takeovers meet that standard. All arguments are implicit [Arg0 Canadian] [Predicate investment] [Arg1 in the US] has declined.

  18. Summary • Nominal SRL system for unstructured text • Supervised argument identification • Works for all 4,700+ distinct NomBank predicates • Argument F1: 75.7% • IAs pose a serious problem for nominal SRL • Performance loss of 8.5% • False positive argument predictions

  19. Research Contributions Investigate implicit arguments (IAs) in nominal SRL • Examine the effect of IAs on nominal SRL performance • Improve SRL performance by taking IAs into account • Recover implicit arguments from discourse

  20. Research Contributions Investigate implicit arguments (IAs) in nominal SRL • Examine the effect of IAs on nominal SRL performance • Improve SRL performance by taking IAs into account • Recover implicit arguments from discourse

  21. Prevalence of Implicit Arguments • Percentage of predicates with local arguments: 57%

  22. Nominal Classification Model • Filter out predicates whose arguments are implicit • Binary classification over token nodes • Positive class: apply standard SRL system • Negative class: ignore

  23. Nominal Classification Model • Features • Primarily syntactic and lexical • Most important feature: ancestor grammar rules S AGR3: S -> NP, VP NP AGR2: VP -> V, NP VP N (John) V (made) NP AGR1: NP -> Det, N Det (a) N (sale)

  24. Nominal Classification Model • Features • Primarily syntactic and lexical • Most important feature: ancestor grammar rules • Other notable features • Candidate token stem • Syntactic category of token’s right sibling • … • Quite different from the SRL argument features

  25. Nominal Classification Evaluation • Training data • Tokens from sections 2-21 of TreeBank • 780,000 training instances • Development data • Section 24 • Feature selection • Threshold tuning • Testing data • Section 23

  26. Nominal Classification Evaluation • Baseline systems • Naïve • All known nouns should be labeled for arguments • MLE • MLE scoring of nouns and a tuned prediction threshold • Results

  27. Nominal Classification Evaluation • Results by nominal distribution

  28. Combined Nominal-Argument Evaluation • Task: identify predicate, then identify arguments Previous best score using ground-truthpredicates: 72.8

  29. Combined Nominal-Argument Evaluation • Results by nominal distribution

  30. Summary • IAs pose a serious problem for nominal SRL • Performance loss of 8.5% • Filter nominal predicates whose arguments are implicit • Predicate F1: 88% • End-to-end argument F1: 71% • Problem: Implicit arguments are not recovered!

  31. Research Contributions Investigate implicit arguments (IAs) in nominal SRL • Examine the effect of IAs on nominal SRL performance • Improve SRL performance by taking IAs into account • Recover implicit arguments from discourse

  32. Research Contributions Investigate implicit arguments (IAs) in nominal SRL • Examine the effect of IAs on nominal SRL performance • Improve SRL performance by taking IAs into account • Recover implicit arguments from discourse

  33. A Motivating Example Georgia-Pacific and Nekoosa produce market pulp, containerboard and white paper. The goods could be manufactured closer to customers, saving shipping costs. • What can traditional SRL systems tell us?

  34. A Motivating Example Georgia-Pacific and Nekoosa produce market pulp, containerboard and white paper. The goods could be manufactured closer to customers, saving shipping costs. • What can traditional SRL systems tell us? • Who is the producer? • What is produced?

  35. A Motivating Example Georgia-Pacific and Nekoosa produce market pulp, containerboard and white paper. The goods could be manufactured closer to customers, saving shipping costs. • What can traditional SRL systems tell us? • Who is the producer? • What is produced? • What is manufactured? • But that’s not the whole story… • Who is the manufacturer?

  36. A Motivating Example Georgia-Pacific and Nekoosa produce market pulp, containerboard and white paper. The goods could be manufactured closer to customers, saving shipping costs. • What can traditional SRL systems tell us? • Who is the producer? • What is produced? • What is manufactured? • But that’s not the whole story… • Who is the manufacturer? • Who ships?

  37. A Motivating Example Georgia-Pacific and Nekoosa produce market pulp, containerboard and white paper. The goods could be manufactured closer to customers, saving shipping costs. • What can traditional SRL systems tell us? • Who is the producer? • What is produced? • What is manufactured? • But that’s not the whole story… • Who is the manufacturer? • Who ships what?

  38. A Motivating Example Georgia-Pacific and Nekoosa produce market pulp, containerboard and white paper. The goods could be manufactured closer to customers, saving shipping costs. • What can traditional SRL systems tell us? • Who is the producer? • What is produced? • What is manufactured? • But that’s not the whole story… • Who is the manufacturer? • Who ships what to whom? Implicit arguments

  39. Implicit Argument Identification • Research questions • Where are implicit arguments? • Can we recover them? • Related work • Japanese anaphora • Indirect anaphora (Sasano et al., 2004) • Zero-anaphora (Imamura et al., 2009) • Implicit arguments • Fine-grained domain model (Palmer et al., 1986) • SemEval Task 10 (Ruppenhofer et al., 2010)

  40. Data Annotation • Ten most prominent NomBank predicates • Derived from verbal role set (ship shipment) • Frequency of nominal predicate • Difference between verbal/nominal argument counts • [John] shipped [the package]. (argument count: 2) • Shipping costs will decline. (argument count: 0) • Predicates instances annotated: 1,247 • Independently annotated by two annotators • Cohen’s Kappa: 64% • Agreement: both unfilled or both filled identically

  41. Annotation Analysis Average number of expressed arguments Post-annotation Verb form Pre-annotation Verb form

  42. Annotation Analysis Average number of expressed arguments Pre-annotation Verb form Post-annotation Verb form

  43. Annotation Analysis • Average arguments across all predicates • Pre-annotation: 1.1 • Post-annotation: 1.9 • Verb form: 2.0 • Overall percentage of possible roles filled • Pre-annotation: 28.0% • Post-annotation: 47.8% ( 71%)

  44. Annotation Analysis 90% within current or previous three sentences Only 55% within current sentence

  45. Model Formulation c1 c2 Georgia-Pacific and Nekoosa produce market pulp, containerboard and white paper. The goods could be manufactured closer to customers, saving shipping costs. c3 • Candidate selection • PropBank/NomBank arguments • Two-sentence candidate window • Coreference chaining • Binary classification function

  46. Model Features SRL structure … … Discourse structure Other

  47. Model Features: SRL Structure • VerbNet role transition Georgia-Pacific and Nekoosa produce market pulp, containerboard and white paper. The goods could be manufactured closer to customers, saving shipping costs. c3

  48. Model Features: SRL Structure • VerbNet role transition Georgia-Pacific and Nekoosa produce market pulp, containerboard and white paper. The goods could be manufactured closer to customers, saving shipping costs. c3 arg1 arg1? create VN role: product product VN class: create VN class: send send item .item • Feature value: create.product send.item • Multiple values are possible

  49. Model Features: SRL Structure • Narrative event chains (Chambers and Jurafsky, 2008) • PMI(manufacture.arg1, ship.arg1) • Computed from SRL output over Gigaword (Graff, 2003) • Advantages: better coverage + relationship strength Georgia-Pacific and Nekoosa produce market pulp, containerboard and white paper. The goods could be manufactured closer to customers, saving shipping costs. c3 arg1

  50. Model Features: Discourse Structure • Penn Discourse TreeBank (Prasad et al., 2008) • Feature value: Contingency.Cause.Result • Might help identify salient discourse segments c2 Georgia-Pacific and Nekoosa produce market pulp, containerboard and white paper. The goods could be manufactured closer to customers, saving shipping costs.

More Related