1 / 14

A Language Modeling Approach to Tracking

A Language Modeling Approach to Tracking. Martijn Spitters & Wessel Kraaij TNO - TPD. TDT2000 at NIST, Gaithersburg. Outline. Introduction: Filtering vs. Tracking Tracking Model Experiments: stemming Influence of normalisation steps merging training stories Conclusions.

jirair
Download Presentation

A Language Modeling Approach to Tracking

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. A Language Modeling Approach to Tracking Martijn Spitters & Wessel Kraaij TNO - TPD TDT2000 at NIST, Gaithersburg

  2. Outline • Introduction: • Filtering vs. Tracking • Tracking Model • Experiments: • stemming • Influence of normalisation steps • merging training stories • Conclusions

  3. Adaptive filtering vs. tracking • Adaptive filtering: • initial short topic statement, relevance judgements available ‘along the way’ • Individual threshold adaptation • Evaluation: Utility function • Tracking: • 1-4 training stories, no other relevance information • uniform threshold • Evaluation: Normalized detection Cost

  4. Changes in system • Model based on P(S|T) instead of P(Q|D) ==>reversed orientation • Uniform threshold requires careful normalization • Likelihood ratio • Story length normalization • Gaussian normalization

  5. TREC8 Adaptive filtering

  6. Tracking Model

  7. Normalization Story length Gaussian:

  8. Experiment 1: Stemming

  9. Experiment 2: Normalization

  10. Experiment 3: Merging Influence of the unbiased training story merging method on four TDT2 topics

  11. TDT2000 Evaluation Results (1)

  12. TDT2000 Evaluation Results (2)

  13. Conclusions • simple language models are effective • Normalisation is important • Stemming is important • Unbiased training story merging procedure seems to help when training story lengths differ substantially

  14. Future plans • Apply EM to learn optimal i • Refine background models • Apply expansion techniques • Unsupervised adaptation

More Related