1 / 24

Backward Machine Transliteration by Learning Phonetic Similarity

Backward Machine Transliteration by Learning Phonetic Similarity. Advisor : Dr. Hsu Presenter : Chien Shing Chen Author: Wei-Hao Lin and Hsin-His Chen. PRESENTED AT SIXTH CONFERENCE ON NATURAL LANGUAGE LEARNING, TAIPEI, TAIWAN,2002. Outline. Motivation Objective Introduction

wind
Download Presentation

Backward Machine Transliteration by Learning Phonetic Similarity

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Backward Machine Transliteration by Learning Phonetic Similarity Advisor :Dr. Hsu Presenter: Chien Shing Chen Author: Wei-Hao Lin and Hsin-His Chen PRESENTED AT SIXTH CONFERENCE ON NATURAL LANGUAGE LEARNING, TAIPEI, TAIWAN,2002

  2. Outline • Motivation • Objective • Introduction • Grapheme-to-Phoneme(音素,音位) Transformation • Similarity Measurement • Learning Phonetic Similarity • Experimental Result • Conclusions • Personal Opinion

  3. Motivation • a similarity-based framework to model the task of backward transliteration • a learning algorithm to automatically acquire phonetic similarities from a corpus • Backward transliteration: from a transliteration to original language, like “本拉登” =>Bin Laden

  4. Objective • Backward machine transliteration by learning phonetic similarity • 雨果(Yu-guo) => Hugo

  5. Introduction

  6. Introduction • IPA : International Phonetic Alphabet(國際音標) • Yu-guo =>h j u g oU • Hugo =>v k uo • Similarity Measurement

  7. Introduction • CMU pronunciation dictionary 0.6 版 • ftp://ftp.cs.cmu.edu/project/fgdata/dict

  8. Similarity Measurement-alignment • Set is the alphabet set of two strings S1 and S2. • ,where ‘_’ stands for space. • Space can be inserted into S1’ and S2’ • S1’ and S2’ are aligned

  9. Similarity Measurement-score • <English,Chinese> <Hugo, Yu3-guo3> • the phoneme pair (v k uo, h j u g oU) • ={h, j, u, v, g, k, oU, uo, _}

  10. Similarity Measurement-score • ={h, j, u, v, g, k, oU, uo, _}

  11. Similarity Measurement-Dynamic • Dynamic programming to trade off : • alignment • similarity scoring matrix M • OPTIMAL • S1 (j h u g oU) • S2 (v k uo)

  12. Dynamic programming-Dynamic • Set T is a n+1 by m+1 table where n is the length S1, m is the length of S2.

  13. Learning Phonetic Similarity • develop a learning algorithm to • remove the efforts of assigning scores in the matrix • capture the subtle difference • How to prepare a training corpus, followed by the learning algorithm.

  14. Learning Phonetic Similarity • Positive pairs: original words and the transliterated words are matched • Negative pairs: mismatch the original words and the transliterated words • Ei: original English • Ci: transliterated Chinese • Corpus with n pairs 克林頓 Clinton 本拉登 Bin Laden 魯賓遜 Robinson n positive pair n (n-1) negative pair

  15. Learning Algorithm • Treat each training sample as a linear equation • m is the size of the phoneme sets, m=9 • wi,jis the row i and the column j of the scoring matrix • xi,j is a binary value indicating the presence of wi,j in the alignment • y is the similarity score.

  16. Learning Algorithm • Linear equation in the corpus can be conveniently represented in the matrix form, • , R is the number of pairs in the corpus • i stands for the ith sample pair in the corpus • wi,jis the scoring matrix • xi,j is a binary value • y is the similarity score

  17. Learning Algorithm • The criterion is the sum-of-squared error minimized. • The classical solution is to take the pseudo inverse of , i.e. ,to obtain the w that minimizes the SSE , i.e. • adopt the Widrow-Hoff rule to solve

  18. Learning Algorithm • k stands for the kth row in the matrix X • i for the number of iterations • is the learning rate • is the momentum coefficient. • is empirically set as as follows,

  19. Learning Algorithm • The w(i) is updated iteratively until the learned w appears to overfit. • The iterations to ensure the w will converge to a vector satisfying • Update w(i) immediately after encountering a new training sample instead of accumulating all errors of training samples • The other speed-up technique is the momentum used to damp the oscillations. .

  20. Experiments • .corpus is consisted of 1574 pairs of <English,Chinese> names • 313 have no entries in the pronouncing dictionary. • 97 phonemes used to represent these names, in which 59 and 51 phonemes are used for Chinese and English names. • Rank is the position of the correct original word in a list of candidate words sorted.

  21. Experiments • .

  22. Experiments • .

  23. Conclusions • Without any phonological analysis, the learning algorithm can acquire those similarities without human intervention.

  24. Personal Opinion • Drawback • obtain the score matrix depend on a few empirically rule • Is the experiment tie in with the testing samples ? • Application • A different method to compute the similarity between words. • Future Work • The Widrow-Hoff rule may estimate the parameter to substitute for attempting intervention blinded. • Combine sound speech recognize with this method to output a new objectivity method

More Related