1 / 24

CS 479, section 1: Natural Language Processing

This work is licensed under a Creative Commons Attribution-Share Alike 3.0 Unported License . CS 479, section 1: Natural Language Processing. Lecture # 36: Alignment and Metrics. Thanks to Dan Klein of UC Berkeley and Uli Germann of ISI for many of the materials used in this lecture.

shanna
Download Presentation

CS 479, section 1: Natural Language Processing

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. This work is licensed under a Creative Commons Attribution-Share Alike 3.0 Unported License. CS 479, section 1:Natural Language Processing Lecture #36: Alignment and Metrics Thanks to Dan Klein of UC Berkeley and Uli Germann of ISI for many of the materials used in this lecture.

  2. Announcements • Check the schedule • Plan enough time to succeed! • Don’t get or stay blocked. • Get your questions answered early. • Get the help you need to keep moving forward. • No late work accepted after the last day of instruction. • Project Report: • Early: Friday • Due: Monday • Reading Report #14 • Phrase-based MT paper • Due: next Wednesday(online again)

  3. Objectives • Consider additional models for statistical word alignment • See how our alignment models capture real phrases • Understand how to actually score a word-level alignment with a word alignment model • Discuss how to evaluate alignment quality

  4. Quiz • Why do we use parallel corpora (bitexts)? • What is the hidden (unknown) variable in building translation models? • What was the main idea behind IBM Model 1? • Model 2?

  5. Recall This Example On Tuesday Nov. 4, earthquakes rocked Japan once again Des tremblements de terre ont à nouveau touché le Japon jeudi 4 novembre. What else is going on here that we haven’t tried to model?

  6. Models Summary • IBM Model 1: word alignment • IBM Model 2: word alignment, with global position (order) model • HMM Model: word alignment, with local position • IBM Model 3: adds model of fertility to model 2, deficient • IBM Model 4: adds relative ordering to model 3, deficient • IBM Model 5: fixes deficiency of Model 4

  7. Context • Given a source language sentence, • The search algorithm must propose possible translations (target language sentences) along with corresponding alignments • Let’s pretend we’re in the midst of a search and scoring a single hypothesis • How do we use our models to compute such a score?

  8. Example: How to score • Spanish source sentence: “Maria no dabaunabofetada a la brujaverde” • Here denotes “foreign” sentence. • During the search, we propose a possible English translation: “Mary did not slap the green witch” • We consider one possible alignment • What is the score, according to model 5? • i.e., what is ?

  9. Example: How to score Examples from Local Models Mary did not slap the green witch n(3|slap) Mary not slap slapslap the green witch P(NULL) Mary not slap slap slap NULL the green witch t(la|the) Maria no dabaunabofetada a la verdebruja d(j|i) Maria no dabaunabofetada a la brujaverde [Al-Onaizan and Knight, 1998]

  10. Example: How to score Mary did not slap the green witch n(3|slap) Mary not slap slapslap the green witch P(NULL) Mary not slap slap slap NULL the green witch t(la|the) Maria no dabaunabofetada a la verdebruja d(j|i) Maria no dabaunabofetada a la brujaverde [Al-Onaizan and Knight, 1998]

  11. Cascaded Training • Standard practice for training: • Initialize one model with the previous (simpler) model • Proceed with EM • Typical order: 1, (2 | HMM), 3, 4, 5

  12. Insight into these models

  13. Examples: Translation and Fertility

  14. Example: Idioms

  15. Example: Morphology

  16. Example: WSD • Word sense disambiguation: • Word-based MT systems rarely have a WSD step • Why not?

  17. Choosing an Alignment

  18. Choosing an Alignment

  19. Efficiency of Choosing Best Alignment

  20. Evaluating TMs • How do we measure TM quality? • Measure quality of the alignments produced • Metric: AER

  21. Alignment Error Rate / Actual

  22. AER • Easy to measure • Problems? • Hard to know what the gold alignments should be • May not correlate with translation quality • like perplexity and speech reco. accuracy in LMs

  23. AER Results [Och & Ney, 2003] Canadian Hansards data

  24. Next • Decoding • Complexity of Decoding • Evaluating Translation

More Related