1 / 28

Dynamic data selection in search (1/4)

A Dynamic In-Search Data Selection Method With Its Application to Acoustic Modeling and Utterance Verification (about training). Dynamic data selection in search (1/4).

gema
Download Presentation

Dynamic data selection in search (1/4)

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. A Dynamic In-Search Data Selection Method With Its Application to Acoustic Modeling and Utterance Verification (about training)

  2. Dynamic data selection in search (1/4) • In continue speech recognition, it become much more difficult to define the CT or TT set because of unknown unit boundaries. • As a result, any possible segmentation in an utterance could potentially become a competing token. However, an exhaustive search it too expensive to be affordable.

  3. Dynamic data selection in search (2/4) • In this paper, every utterance is recognized with Viterbi beam search algorithm. • All partial paths surviving during beam search always have relatively large likelihood values and usually potentially compete with the true path.

  4. Reference phone segmentation alignment Dynamic data selection in search (3/4)

  5. Dynamic data selection in search (4/4) True token of competing Token of

  6. Application I : acoustic modeling (1/20) • When giving true tokes of => MAP • MAP : training observation sequences with prior information

  7. Application I : acoustic modeling (2/20)

  8. A simpler example (DHMM) :

  9. A simpler example (DHMM) :

  10. A simpler example (DHMM) :

  11. A simpler example (DHMM) :

  12. Application I : acoustic modeling (3/20)

  13. Application I : acoustic modeling (4/20)

  14. Application I : acoustic modeling (5/20)

  15. Application I : acoustic modeling (6/20)

  16. Application I : acoustic modeling (7/20)

  17. Application I : acoustic modeling (8/20) (2) (3) (1)

  18. Application I : acoustic modeling (9/20)

  19. Application I : acoustic modeling (10/20)

  20. Application I : acoustic modeling (11/20) • When giving true tokes of => MCE • Imposter word : the hypothesized word is wrong, and the likelihood exceeds the likelihood given its reference model. • If we can minimize the total number of imposter word, we can reduce the WER.

  21. Application I : acoustic modeling (12/20) MCE • The misclassification distance measure for the wrong word W is defined as • And then a general form of the “smoothed” count of the imposter word is defined as

  22. Application I : acoustic modeling (13/20) MCE • The so-called GPD algorithm is adopted to minimized the “smoothed” count of the imposter word

  23. Application I : acoustic modeling (14/20) MCE

  24. Application I : acoustic modeling (15/20)

  25. Application I : acoustic modeling (16/20) Reference model

  26. Application I : acoustic modeling (17/20) Reference model

  27. Application I : acoustic modeling (18/20)

More Related