1 / 24

CRM114 vs. Mr X: Notes for the NIST TREC 2005 Spam Track

CRM114 vs. Mr X: Notes for the NIST TREC 2005 Spam Track. William S. Yerazunis* presenting for the CRM114 Team TREC 2005 Closing Plenary ( It should be noted that CRM114 is GPLed open source, and flourishes under the benevolent neglect of numerous companies and universities. )

tierra
Download Presentation

CRM114 vs. Mr X: Notes for the NIST TREC 2005 Spam Track

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. CRM114 vs. Mr X: Notes for the NIST TREC 2005 Spam Track William S. Yerazunis* presenting for the CRM114 Team TREC 2005 Closing Plenary ( It should be noted that CRM114 is GPLed open source, and flourishes under the benevolent neglect of numerous companies and universities. ) * MERL – Mitsubishi Electric Research Laboratories

  2. The Real Goal • The real goal is not to “get the best numbers” • The real goal is to destroy the spammer’s business model. • Remember that when you make your engineering trade-off decisions....

  3. The Real Goal • Accept that filtering might not be the only thing you need. Consider: • Authentication • Shared blacklists • Server or source signatures • Immunization • That said, what can we learn?

  4. If you look at numbers for the 44 filters tested at TREC: • Of the 44 filter configurations, CRM114 is typically in the lowest 1-ROCAC% group (1-ROCAC% is the probability of a random spam scoring better than a random good email, so low is better. 1-ROCAC% is not unreasonable as a spam filter figure of merit) • Of the eight “sweet spots” (error rates when operated with a fixed 1% error in the opposite class) CRM114 is always either best, or statistically indistinguishable from the best filter tested.

  5. So, CRM114 did Really Well. What does CRM114 do that’s different? What does CRM114 do that’s the same? What part of that is “portable”? ..... “What’s in the Secret Sauce?”

  6. What Does CRM114 Do Differently? • Speed matters! • Don’t throw away information • Let the computer do the hard parts • Openness matters- open source, open mind

  7. Speed Matters! • Whatever you do, think about the impact of your new gem of coding. • If your filter is too slow, it will never get wide deployment, which means it won’t impact the spam business model and thus, the spam filter fails in its real goal.

  8. Speed Matters! • Whatever you do, think about the impact of your new gem of coding. • If your filter is too slow, it will never get wide deployment, which means it won’t impact the spam business model and thus, the spam filter fails in its real goal. • A slow filter also means you won’t be able to test many variations of the filter.

  9. Information Theory... • Think of the whole problem on an information-theoretic basis. • That is, it's not just bits, but bits-in-context • Avoid throwing away information.

  10. Information Theory... Avoid throwing away information • Unlike most Graham-style filters, CRM114 does not use a “significance window” of the most extreme N words. Every feature counts. • This means no word or feature has an overriding impact. There’s no “ten nonspammy words” that can sneak a spam past the filter.

  11. Information Theory... Avoid throwing away information • With no top-N extremum values, CRM114 can use a very gentle conditional probability formula, so statistical outlier features (statistical noise) have low impact. • CRM114’s conditional probabilities are limited to roughly the range [ .47 ... .53 ] at 1:1 [ .44 ... .56 ] at 10:1 [ .43 ... .57 ] at 1000:1

  12. Information Theory... Bits In Context: • Tuple-based features (such as OSB) are much better than single-word features. Example: the string “foo bar baz wugga” yields: • foo bar • foo baz • foo wugga • bar baz, • bar wugga

  13. Information Theory... Let the computer do the hard work • Tuple-based features are much more robust than ad-hoc filters like “__header:” • Don’t bother trying to guess heuristics. That’s too much like work. • Computers are good at accounting- so let the computer figure out what’s significant. Humans have better things to do with their time.

  14. Information Theory... Observation: If your feature set is rich, you can use just about any combining rule or database and get publishable results.

  15. Speed Matters!!! • How you store your statistics matters (from the engineering point of view) • Special-purpose hashing systems can be much faster than general-purpose databases • Not just a little faster – hundreds of times faster. • This really matters when you are doing testing.

  16. Test like crazy. • Have multiple test corpora. • There are huge disparities across corpora • The No Free Lunch theorem (Wolpert and Macready, 1997) hits with a vengeance here. • Expect 2:1 or worse disparities on an everyday basis, and 10:1 disparities on a monthly basis. • If you don’t test, you won’t know. And if testing is too slow, you won’t test. • Don’t be afraid to try different configurations

  17. Speed Matters!!! • You will make far better progress if you can test against a 5,000 message corpus in 10 minutes than if you have to take a weekend to let your test run. • The fastest CRM114 configurations are capable of running the entire 4147 message (2003 SpamAssassin corpus) in under 1 minute!

  18. Stop-words • Don'tdo "stop-word elimination". • Stop words may be common, but they still carry information, especially in a tuple-feature system. • Remember: it’s not bits – it’s bits in context.

  19. Symmetry ? • Design Decision: Should a filter err on the side of “let it in”? • OPINION: A good text filter should be completely symmetrical (at least by default). • Hypothesis: a falsely “accepted” spam can be as costly as a falsely “rejected” legitimate mail. • Substantiation: consider the cost of a bank-authentication spam that your Grandma “falls for”...

  20. Hapaxes and Grooming • Hapaxes carry information- but that information only becomes useful when the hapax is seen again (and thus proven to not really be a hapax). • Your database may well have “hidden channel” information on features such as time-last-seen. • Even if it’s only relative, or approximate, this information is very useful for getting rid of the stale database entries first.

  21. No Free Lunches – Part 2 • The No Free Lunch theorem proves that there is no one "best" classifier. • Empirical observation: There is no one best classifier author, either.

  22. No Free Lunches – Part 2 • The No Free Lunch theorem proves that there is no one "best" classifier. • Empirical observation: There is no one best classifier author, either. • If you make your architecture open and pluggable, then good filter people will write code for you, and you can take the credit!*

  23. *So, in the interests of honesty... ...I need to give very strong acknowledgments and HUGE thanks to: Fidelis Assis, Shalendra Chhabra, Jaakko Hyvätti, Barry Jaspan, Ville Laurikari, Raul Miller, Paolo Panizza, Christian Siefkes, and the hundreds of code-readers, bug-whackers, and other-wise contributors.

  24. Summary • Speed matters. • Don’t throw away information. • Use tuples. Keep stopwords. Keep hapaxes. • Be symmetrical. • Benchmark! Benchmark! Benchmark! • Be open. Open source, open mind. • People want to help. Help them to help. • ... AND ACKNOWLEDGE THEM! Thank you very much. Are there any questions?

More Related