1 / 25

Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews

Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews. Peter D. Turney Institute for Information Technology National Research Council of Canada Ottawa, Ontario, Canada, K1A 0R6 peter.turney@nrc.ca

clancy
Download Presentation

Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification on Reviews Peter D. Turney Institute for Information Technology National Research Council of Canada Ottawa, Ontario, Canada, K1A 0R6 peter.turney@nrc.ca Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL), Philadelphia, July 2002, pp. 417-424.

  2. 1. Introduction • If you are considering a vacation in Akumal, Mexico, you might go to a search engine and enter the query “Akumal travel review”. Google reports about 5,000 matches. It would be useful to know what fraction of these matches recommend Akumal as a travel destination • Other application: Recognizing flames, Developing new kinds of search tools

  3. 1. Introduction • This paper present a simple unsupervised learning algorithm for classifying a review as recommended or not recommended • Input: written review, Output: classification • Using POS tagger to identify phrases in the input text that contain adjectives or adverbs • Estimating the semantic orientation of each extracted phrase • Assigning the given review to a class, recommended or not recommended, based on the average semantic orientation of the phrases extracted from the review

  4. 1. Introduction • The PMI-IR algorithm is employed to estimate the semantic orientation of a phrase (Turney, 2001) • PMI-IR uses Pointwise Mutual Information (PMI) and Information Retrieval (IR) to measure the similarity of pairs of words or phrases. • The semantic orientation of a given phrase is calculated by comparing its similarity to a positive reference word “excellent” with its similarity to a negative reference word “poor”

  5. 2. Classifying Reviews • Past work has demonstrated that adjectives are good indicators of subjective, evaluative sentences (Hatzivassiloglou & Wiebe, 2000; Wiebe, 2000; Wiebe et al., 2001). • However, although an isolated adjective may indicate subjectivity, there may be insufficient context to determine semantic orientation. • Ex: unpredictable, simple etc.

  6. 2. Classifying Reviews • The first step a POS tagger is applied to the review (Brill, 1994) Two consecutive words are extracted from the review if their tags conform to any of the patterns

  7. 2. Classifying Reviews • The second step is to estimate the semantic orientation of the extracted phrases, using the PMI-IR algorithm. This algorithm uses mutual information as a measure of the strength of semantic association between two words (Church & Hanks, 1989).

  8. 2. Classifying Reviews • The Pointwise Mutual Information (PMI) between two words, word1 and word2, is defined as follows (Church & Hanks, 1989): • SO(phrase) = PMI(phrase, “excellent”) – PMI(phrase, “poor”) (2)

  9. 2. Classifying Reviews • PMI-IR estimates PMI by issuing queries to a search engine and noting the number of hits (matching documents). The following experiments use the AltaVista Advanced Search engine, which indexes approximately 350 million web pages • Choosing AltaVista because it has a NEAR operator. The AltaVista NEAR operator constrains the search to documents that contain the words within ten words of one another, in either order.

  10. 2. Classifying Reviews • To avoid division by zero, adding 0.01 to the hits • Also skipped phrase when both hits(phrase Near “excellent”) and hits(phrase Near “poor”) were less than four

  11. 2. Classifying Reviews • The third step is to calculate the average semantic orientation of the phrases in the given review and classify the review as recommended if the average is positive and otherwise not recommended

  12. 2. Classifying Reviews

  13. 2. Classifying Reviews

  14. 3. Related Work • This work is most closely related to Hatzivassiloglou and McKeown’s (1997) work on predicting the semantic orientation of adjectives. They note that there are linguistic constraints on the semantic orientations of adjectives in conjunctions

  15. 3. Related Work • The tax proposal was simple and well received by the public. • The tax proposal was simplistic but well received by the public. • (*) The tax proposal was simplistic and well received by the public.

  16. 3. Related Work • All conjunctions of adjectives are extracted from the given corpus. • A supervised learning algorithm combines multiple sources of evidence to label pairs of adjectives as having the same semantic orientation or different semantic orientations. The result is a graph where the nodes are adjectives and links indicate sameness or difference of semantic orientation.

  17. 3. Related Work • A clustering algorithm processes the graph structure to produce two subsets of adjectives, such that links across the two subsets are mainly different-orientation links, and links inside a subset are mainly same-orientation links • Since it is known that positive adjectives tend to be used more frequently than negative adjectives, the cluster with the higher average frequency is classified as having positive semantic orientation. • This algorithm classifies adjectives with accuracies ranging from 78% to 92%, depending on the amount of training data that is available.

  18. 3. Related Work • Other related work is concerned with determining subjectivity (Hatzivassiloglou & Wiebe, 2000; Wiebe, 2000; Wiebe et al., 2001). • The task is to distinguish sentences that present opinions and evaluations from sentences that objectively present factual information (Wiebe, 2000) • A variety of potential applications for automated subjectivity tagging, such as recognizing “flames” (Spertus, 1997), classifying email, recognizing speaker role in radio broadcasts, and mining reviews.

  19. 4. Experiments

  20. 4. Experiments

  21. 5. Discussion of Results

  22. 5. Discussion of Results

  23. 5. Discussion of Results • A limitation of PMI-IR is the time required to send queries to AltaVista. Inspection of Equation (3) shows that it takes four queries to calculate the semantic orientation of a phrase

  24. 6. Applications • Providing summary statistics for search engines. Given the query “Akumal travel review”, a search engine could report, “There are 5,000 hits, of which 80% are thumbs up and 20% are thumbs down.” • Filtering “flames” for newsgroups (Spertus, 1997).

  25. 7. Conclusions • Movie reviews are difficult to classify, because the whole is not necessarily the sum of the parts • On the other hand, for banks and automobiles, it seems that the whole is the sum of the parts • The simplicity of PMI-IR may encourage further work with semantic orientation. • The limitations of this work include the time required for queries and, for some applications, the level of accuracy that was achieved.

More Related