1 / 19

An Overview of Text Mining Rebecca Hwa 4/25/2002

References M. Hearst, “ Untangling Text Data Mining ,” in the Proceedings of the 37 th Annual Meeting of the Association for Computational Linguistics , 1999.

Jimmy
Download Presentation

An Overview of Text Mining Rebecca Hwa 4/25/2002

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. References M. Hearst, “Untangling Text Data Mining,” in the Proceedings of the 37th Annual Meeting of the Association for Computational Linguistics, 1999. E. Riloff and R. Jones, “Learning Dictionaries for Information Extraction Using Multi-level Boot-strapping,” in the Proceedings of AAAI-99, 1999. K. Nigam, A. McCallum, S. Thrun, and T. Mitchell, “Text Classification from Labeled and Unlabeled Documents using EM,” in Machine Learning, 2000. M. Grobelnik, D. Mladenic, and N. Milic-Frayling, “Text Mining as Integration of Several Related Research Areas: Report on KDD’2000 Workshop on Text Mining,” 2000. An Overview of Text MiningRebecca Hwa4/25/2002

  2. What Is Text Mining? “The objective of Text Mining is to exploit information contained in textual documents in various ways, including …discovery of patterns and trends in data, associations among entities, predictive rules, etc.” (Grobelnik et al., 2001) “Another way to view text data mining is as a process of exploratory data analysis that leads to heretofore unknown information, or to answers for questions for which the answer is not currently known.” (Hearst, 1999)

  3. Text Mining • How does it relate to data mining in general? • How does it relate to computational linguistics? • How does it relate to information retrieval?

  4. Challenges in Text Mining • Data collection is “free text” • Data is not well-organized • Semi-structured or unstructured • Natural language text contains ambiguities on many levels • Lexical, syntactic, semantic, and pragmatic • Learning techniques for processing text typically need annotated training examples • Consider bootstrapping techniques

  5. Text Mining Tasks • Exploratory Data Analysis • Using text to form hypotheses about diseases (Swanson and Smalheiser, 1997). • Information Extraction • (Semi)automatically create (domain specific) knowledge bases, and then use standard data-mining techniques. • Bootstrapping methods (Riloff and Jones, 1999). • Text Classification • Useful intermediary step for information extraction • Bootstrapping method using EM (Nigam et al., 2000).

  6. Biomedical Data Exploration (Swanson, and Smalheiser, 1997) • Extract pieces of evidence from article titles in the biomedical literature • “stress is associated with migraines” • “stress can lead to loss of magnesium” • “calcium channel blockers prevent some migraines” • “magnesium is a natural calcium channel blocker” • Induce a new hypothesis not in the literature by combining culled text fragments with human medical expertise • Magnesium deficiency may play a role in some kinds of migraine headache

  7. Challenges in Data Exploration • How can valid inference links be found without succumbing to combinatorial explosion of possibilities? • Need better models of lexical relationships and semantic constraints (very hard) • How should the information be presented to the human experts to facilitate their exploration?

  8. Information Extraction (IE) • Extract domain-specific information from natural language text • Need a dictionary of extraction patterns (e.g., “traveled to <x>” or “presidents of <x>”) • Constructed by hand • Automatically learned from hand-annotated training data • Need a semantic lexicon (dictionary of words with semantic category labels) • Typically constructed by hand

  9. Challenges in IE • Automatic learning methods are typically supervised (i.e., need labeled examples) • But annotating training data is a time-consuming and expensive task. • Can we develop better unsupervised algorithm? • Can we make better use of a small set of labeled example?

  10. Learning Dictionaries for IE via Bootstrapping (Riloff and Jones, 1999) • Simultaneously learn extraction patterns and domain-specific semantic lexicons • Input requires a small set of seed words (for the semantic categories) and a large collection of text • Mutual bootstrapping • Learns extraction patterns from seed words • Use extraction patterns to identify new words to add to the semantic categories • Meta-bootstrapping to reduce noise

  11. Text classification (TC) • Tag a document as belonging to one of a set of pre-defined classes • “This does not lead to discovery of new information…” (Hearst, 1999). • Many practical uses • Group documents into different domains (useful for domain specific information extraction) • Learn reading interests of users • Automatically sort e-mail • On-line New Event Detection

  12. Challenges in TC • Like IE, also need lots of labeled examples as training data • After a user has labeled 1000 UseNet news articles, the system was only right ~50% of the time at selecting articles interesting to the user. • What other sources of information can reduce the need for labeled examples?

  13. TC from Labeled and Unlabeled Documents using EM (Nigam et al., 2000) • Expectation-Maximization • Iterative algorithm for MLE in parametric estimation problems with missing data (e.g. the labels for the example) • Nigam et al. combined the EM algorithm with a Naïve Bayes classifier, using both labeled and unlabeled data as input • Dynamically adjust strength of unlabeled data’s contribution to parameter estimation in EM • Reduce the bias of naïve Bayes by modeling each class with multiple mixture components

  14. Probabilistic Framework for TC • Assumption #1: Doc produced by mixture model • Generate docs according to probability distribution defined by the model parameters q • Assumption #2: Each class is modeled by one mixture component: C ={c1,…,c|C|} Prob. of model generating doc di is:

  15. Naïve Bayes Model • Assumes words in the document are generated independently (no context) • Assume all text have the same length • Model parameters:

  16. Using a Trained Model • What class should a new document d be assigned to? • Pick the class with the highest probability

  17. Parameter Estimation with Labeled Documents • Estimating model parameters:

  18. Parameter Estimation with Unlabeled Documents • EM: for “incomplete data” problems • Maximize prob. of model generating observed data • Build initial classifier (initialize the parameters to “reasonable” starting values) • Repeat until convergence • E-Step: Use current classifier params, qt, to estimate P(c|d;qt) for all d in Du • M-Step: Re-estimate the classifier, qt+1, using the expected counts from the E-Step

  19. Augmented EM • Weight the unlabeled data • Otherwise, unlabeled data overwhelms the small amount of labeled data • Modify M-step to multiply expected counts with a weight factor • Relax the one class one mixture component assumption • Allow labeled data to fall into “topics” within a class • Modify E-step to allow labeled document to probabilistically belong to sub-topics

More Related