1 / 27

PROBABILISTIC CFGs & PROBABILISTIC PARSING - PowerPoint PPT Presentation

PROBABILISTIC CFGs & PROBABILISTIC PARSING. Universita’ di Venezia 3 Ottobre 2003. Probabilistic CFGs. Context-Free Grammar Rules are of the form: S  NP VP In a Probabilistic CFG, we assign a probability to these rules: S  NP VP, P(SNP,VP|S). Why PCFGs?.

I am the owner, or an agent authorized to act on behalf of the owner, of the copyrighted work described.

PowerPoint Slideshow about ' PROBABILISTIC CFGs & PROBABILISTIC PARSING ' - bishop

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.

- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -
Presentation Transcript

PROBABILISTIC CFGs &PROBABILISTIC PARSING

Universita’ di Venezia

3 Ottobre 2003

• Context-Free Grammar Rules are of the form:

• S  NP VP

• In a Probabilistic CFG, we assign a probability to these rules:

• S  NP VP, P(SNP,VP|S)

DISAMBIGUATION: with a PCFG, probabilities can be used to choose the most likely parse

ROBUSTNESS: rather than excluding things, a PCFG may assign them a very low probability

LEARNING: CFGs cannot be learned from positive data only

PCFGs in Prolog (courtesy Doug Arnold)

s(P0, [s,NP,VP] ) --> np(P1,NP), vp(P2,VP), { P0 is 1.0*P1*P2 }.

….vp(P0, [vp,V,NP] ) --> v(P1,V), np(P2,NP ), { P0 is 0.7*P1*P2 }.

PCFGs specify a language model, just like n-grams

We need however to make some independence assumptions yet again: the probability of a subtree is independent of:

Using PCFGs to disambiguate: “Astronomers saw stars with ears”

Parsing with PCFGs:A comparison with HMMs

An HMM defines a REGULAR GRAMMAR:

Parsing with CFGs: A comparison with HMMs

Inside and outside probabilities(cfr. forward and backward probabilities for HMMs)

Reconstruct the rules used in the analysis of the Treebank

Estimate probabilities by:P(AB) = C(AB) / C(A)

Probabilistic lexicalised PCFGs(Collins, 1997; Charniak, 2000)

• Manning and Schütze, chapters 11 and 12

• Some slides and the Prolog code are borrowed from Doug Arnold

• Thanks also to Chris Manning & Diego Molla