1 / 27

Frank Reichartz Gerhard Paaß Knowledge Discovery Fraunhofer IAIS St. Augustin, Germany

Estimating Supersenses with Conditional Random Fields. Frank Reichartz Gerhard Paaß Knowledge Discovery Fraunhofer IAIS St. Augustin, Germany. Agenda. Agenda. Introduction Models for Supersenses Conditional Random Fields Lumped Observations Summary. è. Use Case Contentus.

maina
Download Presentation

Frank Reichartz Gerhard Paaß Knowledge Discovery Fraunhofer IAIS St. Augustin, Germany

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Estimating Supersenses with Conditional Random Fields Frank Reichartz Gerhard Paaß Knowledge Discovery Fraunhofer IAIS St. Augustin, Germany

  2. Agenda Agenda • Introduction • Models for Supersenses • Conditional Random Fields • Lumped Observations • Summary è

  3. Use Case Contentus • Digitize a Multimedia Collection of German National Library • Music of former GDR • Digitize • Quality control • Meta data collection • Semantic indexing • Semantic search engine • Target • Provide content: text, score sheets, video, images, speech • Generate meta data: composers, premiere, director, artists, … • Extract entities: dates, places, relations, composers, pieces of music, … • Assign meanings to words and phrases: use ontology

  4. Wordnet as Ontology • WordNet is a fine-grained word sense hierarchy • The same word may have different senses: bank = financial institutebank = river boundary • Defines senses (synsets) for • Verbs • Common & proper nouns • Adjectives • Adverbs Target: assign each word to a synset • Easy semantic indexing & retrieval

  5. Fine Grained Word Senses • Example: senses of noun „blow“ Very subtle differences between senses

  6. Hierarchy of Hypernyms • Supersense level • Fewer distinctions • Retains main differences Target: assign verbs / nouns to a supersense

  7. List of Supersenses 26 15

  8. Supersenses discriminate between many synsets Noun blow: • 7 synsets • 5 supersenses Verb blow: • 22 synsets • 9 supersenses Sufficient for coarse disambiguation

  9. Agenda Agenda • Introduction • Models for Supersenses • Conditional Random Fields • Lumped Observations • Summary è

  10. Training Data: SemCor Dataset Synset Supersense A A DT compromise compromise NN 1190419 noun.communication will will MD leave leave VB 2610151 verb.stative both both DT sides side NN 8294366 noun.group without without IN the the DT glow glow NN 13864852 noun.state of of IN triumph triumph NN 7425691 noun.feeling , , PUNC but but CC it it PRP will will MD save save VB 2526596 verb.social Berlin location NNP 26074 noun.location . . PUNC Input Output

  11. Prior Work: Classifier • Bag-of-words is not sufficient: code relative positions • Use classifiers • MaxEnt • SVM • Naive Bayes • kNN • Proc. SemEval 2007Coarse-Grained English All-Words Task

  12. Prior Work: Sequence Modelling Chiaramita & Altun 06: Use perceptron-trained HMM • Maximize predictive performance on training set • Ignore ambiguity: use only most frequent sense Deschacht & Moens 06: Use Conditional Random Field • Exploit hierarchy to model many classes • Apply to fine grained word sense: good results

  13. Agenda Agenda • Introduction • Models for Supersenses • Conditional Random Fields • Lumped Observations • Summary è

  14. Definition of Conditional Random Fields • observed words / features X1,…,Xn • states Y1,…,Yn • each state Yt may be influenced by many of the X1,…,Xnè features Definition: Let G=(V,E) be a graph such that Y=(Yt)tÎV, so that Y is indexed by the vertices of G. Then (X,Y) is a conditional random field in case, when conditioned on X, the random variables Yt obey the Markov property with respect to the graph: • Hammersley Cliffort Theorem: probability can be written as a product of potential functions • Variable sets are cliques in the neighborhood set

  15. Simplification: Sequential Chain • observed words / features X1,…,Xn • states Y1,…,Yn • features may involve two consecutive states and all observed words examples • feature has value 1 if Yt-1="other" and Yt=location and Xthas pos tag "proper name" and (Xt-2,Xt-1)="arrived in". Otherwise the value is 0 • estimated POS tags, noun phrase tags, weekday, amounts, etc. • prefixes, suffixes; matching regular expressions for capitalization, etc. • information from lexica, lists of proper names [Lafferty, McCallum, Pereira 01]

  16. Derivative of Likelihood estimate the optimal parameter vector l for the training set (Xd,Yd), d=1,…,D observed feature value expected feature value how can we calculate the expected feature values? • need for every document d and state Yd,t the probability p(Ydt=i|Xd) • need for every d and states Yd,t , Yd,t+1 the probability p(Yd,t=i,Yd,t+1=j|Xd) use forward - backward algorithm as for the hidden Markov model

  17. Optimization Procedure: Gradient Ascent observed feature value expected feature value regularization • Regularization term: use small weight values, if possible è smaller generalization errorBayesian prior distribution: Gaussian, Laplace, etc. • use gradient-based optimizer: e.g. conjugate gradient, BFGS è approximate quadratic optimization • use stochastic gradient [Sutton, McCallum 06]

  18. Features • Lemmas of verbs of previous position • Part-of-Speech of lags -2|-1|0|1|2 • Coarse POS lags -2|-1|0|1|2 • Three letter prefixes lag -1,0,1 • Three letter suffixes lag -1,0,1 • INITCAP -1|0|1 • ALLDIGITS -1|0|1 • ALLCAPS -1|0|1 • MIXEDCAPS -1|0|1 • CONTAINSDASH -1|0|1 • Class-ID of unsupervised LDA topic model with 50 classes SemCor training set: ~ 20k sentences, 5-fold cross validation

  19. Noun Results for nouns • Different F1-valuesevent: 67.0%Tops: 98.2% • Micro-Average: 83.5% • Macro-Average: 77.9% • Different frequencies of examplesmotive: 133artifact: 8894

  20. Comparison to Prior Result Micro-Average: • Ciaramita & Altun 06: 77.18 % (s=0.45) • CRF 83.5% (s=0.11)~ 28% reduction of error

  21. Agenda • Introduction • Models for Supersenses • Conditional Random Fields • Lumped Observations • Summary è

  22. Need for More Data • WordNet covers more than 100000 synsets • Few examples per supersense: higher training error • Many examples required to train each synsetSemCor: ~20k sentences • Manual labelling is costly Exploit restrictions in WordNet • Each word has only a subset of possible supersenses • Blow: n.act, n.event, n.phenomenon, n.artifact, n.act • Unlabeled data: assign possible supersenses to each word Specialized CRF required

  23. Conditional Random Field with Lumped Supersenses • observed words / features X1,…,Xn • states Y1,…,Yn, YtÎ{h1,…, hk} • Observations: Subsets of supersenses Yt Í{h1,…, hk} • An observation (X1,…,Xn; Y1,…,Yn) contains a large number of sequences • Adapt likelihood computation

  24. Training Data: SemCor Dataset Possible Supersenses A A DT compromise compromise NN n.communication, n.act will will MD leave leave VB v.stative, v.motion, v.cogn., v.change, v.social, v.possession both both DT sides side NN n.group, n. location, n.body, n.artifact, n.cogn., n. food, n. communication, n.object, n.event without without IN the the DT glow glow NN n.state, n.attribute, n.phenomenon, n.feeling of of IN triumph triumph NN n.feeling, n.event , , PUNC but but CC it it PRP will will MD save save VB v.social, v.possession, v.change, Berlin location NNP n.location, n.person, n.artifact . . PUNC

  25. Results for Lumped Supersenses: SemCor Work in progress • Simulate lumped supersenses • Determine possible supersenses for SemCor • Use different fractions of annotated / possible supersenses • Supersenses estimated without annotations: only 0.8%reduction of F-value

  26. Agenda • Theseus Overview • Use Case Contentus • Core Technology Cluster • Supersense Tagging • Summary & Conclusions è

  27. Summary • Sequence models are able to extract supersenses • New features like topic models help • We may use non-annotated texts by exploiting restrictions in the ontology • Chance to improve classifiers considerably • May enhance higher order IE and information retrieval Todo • Apply to lower levels of hierarchy • Detect new senses / supersenses of words in WordNet

More Related