1 / 22

Using CTW as a language modeler in Dasher

Using CTW as a language modeler in Dasher. Martijn van Veen 05-02-2007 Signal Processing Group Department of Electrical Engineering Eindhoven University of Technology. Overview. What is Dasher And what is a language model What is CTW And how to implement it in Dasher

huslu
Download Presentation

Using CTW as a language modeler in Dasher

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Using CTW as a language modeler in Dasher Martijn van Veen 05-02-2007 Signal Processing Group Department of Electrical Engineering Eindhoven University of Technology

  2. Overview • What is Dasher • And what is a language model • What is CTW • And how to implement it in Dasher • Decreasing the model costs • Conclusions and future work

  3. Dasher • Text input method • Continuous gestures • Language model • Let’s give it a try! Dasher

  4. Dasher: Language Model • Conditional probability for each alphabet symbol, given the previous symbols • Similar to compression methods • Requirements: • Sequential • Fast • Adaptive • Model is trained • Better compression -> faster text input

  5. Dasher: Language model • PPM: Prediction by Partial Match • Predictions by models of different order • Weight factor for each model

  6. Dasher: Language model • Asymptotically PPM reduces to fixed order context model • But the incomplete model works better!

  7. CTW: Tree model • Source structure in the model, parameters memoryless • KT estimator: a = number of zeros b = number of ones

  8. CTW: Context tree • Context-Tree Weighting: combine all possible tree models up to a maximum depth

  9. CTW: tree update

  10. CTW: Implementation • Current implementation • Ratio of block probabilities stored in each node • Efficient but patented • Develop a new implementation • Use only integer arithmetic, avoid divisions • Represent both block probabilities as fractions • Ensure denominators equal by cross-multiplication • Store the numerators, scale if necessary

  11. CTW for Text • Binary decomposition • Adjust zero-order estimator

  12. Results • Comparing PPM and CTW language models • Single file • Model trained with English text • Model trained with English text and user input

  13. CTW: Model costs • What are model costs?

  14. CTW: Model costs • Actual model and alphabet size fixed -> Optimize weight factor alpha • Per tree -> not enough parameters • Per node -> not enough adaptivity • Optimize alpha per depth of the tree

  15. CTW: Model costs • Exclusion: only use Betas of the actual model • Iterative process • Convergent? • Approximation: To find actual model use Alpha = 0.5

  16. CTW: Model costs • Compression of an input sequence • Model costs significant, especially for short sequence • No decrease by optimizing alpha per depth?

  17. CTW: Model costs • Maximize probability in the root, instead of the probability per depth • Exclusion based on alpha = 0.5 almost optimal

  18. CTW: Model costs Results in Dasher scenario: • Trained model • Negative effect if no user text is available • Trained with concatenated user text • Small positive effect if user text added to training text, and very similar to it

  19. Conclusions • New CTW Implementation • Only integer arithmetic • Avoids patented techniques • New decomposition tree structure • Dasher language model based on CTW • 6 percent more are accurate predictions than PPM-D • Decreasing the model costs • Only insignificant decrease possible with our method

  20. Future work • Make CTW suitable for MobileDasher • Decrease memory usage • Decrease number of computations • Combine language models • Select locally best model, or weight models together • Combine languages in 1 model • Models differ in structure or in parameters?

  21. Thank you for your attention Ask away!

  22. CTW: Implementation • Store the numerators of the block probabilities

More Related