1 / 61

Lecture 2: The term vocabulary and postings lists Related to Chapter 2:

Lecture 2: The term vocabulary and postings lists Related to Chapter 2: http://nlp.stanford.edu/IR-book/pdf/02voc.pdf. Ch. 1. Recap of the previous lecture. Basic inverted indexes: Structure: Dictionary and Postings Key step in construction: Sorting Boolean query processing

koen
Download Presentation

Lecture 2: The term vocabulary and postings lists Related to Chapter 2:

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Lecture 2: The term vocabulary and postings lists Related to Chapter 2: http://nlp.stanford.edu/IR-book/pdf/02voc.pdf

  2. Ch. 1 Recap of the previous lecture • Basic inverted indexes: • Structure: Dictionary and Postings • Key step in construction: Sorting • Boolean query processing • Intersection by linear time “merging” • Simple optimizations

  3. Tokenizer Friends Romans Countrymen Token stream. Linguistic modules friend Modified tokens. friend roman countryman roman Indexer 2 4 countryman 1 2 Inverted index. 16 13 Recall the basicindexing pipeline Documents to be indexed. Friends, Romans, countrymen.

  4. Plan for this lecture Elaborate basic indexing • Preprocessing to form the term vocabulary • Documents • Tokenization • What terms do we put in the index? • Postings • Faster merges: skip lists • Positional postings and phrase queries

  5. Sec. 2.1 Parsing a document • What format is it in? • pdf/word/excel/html? • What language is it in? • What character set is in use? Each of these is a classification problem, which we will study later in the course. But these tasks are often done heuristically …

  6. Sec. 2.1 Complications: Format/language • Documents being indexed can include docs from many different languages • A single index may have to contain terms of several languages. • Sometimes a document or its components can contain multiple languages/formats • French email with a German pdf attachment. • Document unit

  7. Tokens and Terms

  8. Tokenization • Given a character sequence and a defined document unit, tokenization is the task of chopping it up into pieces, called tokens, perhaps at the same time throwing away certain characters, such as punctuation.

  9. Sec. 2.2.1 Tokenization • Input: “Friends, Romans, Countrymen” • Output: Tokens • Friends • Romans • Countrymen • A token is a sequence of characters in a document • Each such token is now a candidate for an index entry, after further processing • Described below • But what are valid tokens to emit?

  10. Sec. 2.2.1 Issues in tokenization • Iran’s capital  Iran? Irans? Iran’s? • Hyphen • Hewlett-Packard  Hewlett and Packard as two tokens? • the hold-him-back-and-drag-him-away maneuver: break up hyphenated sequence. • co-education • lowercase, lower-case, lower case ? • Space • San Francisco: How do you decide it is one token?

  11. Sec. 2.2.1 Issues in tokenization • Numbers • Older IR systems may not index numbers • But often very useful: think about things like looking up error codes/stack traces on the web • (One answer is using n-grams: Lecture 3) • Will often index “meta-data” separately • Creation date, format, etc. • 3/12/91 Mar. 12, 1991 12/3/91 • 55 B.C. • B-52 • My PGP key is 324a3df234cb23e • (800) 234-2333

  12. Sec. 2.2.1 Language issues in tokenization • French • L'ensemble one token or two? • L ? L’ ? Le ? • Want l’ensemble to match with un ensemble • Until at least 2003, it didn’t on Google • German noun compounds are not segmented • Lebensversicherungsgesellschaftsangestellter • ‘life insurance company employee’ • German retrieval systems benefit greatly from a compound splitter module • Can give a 15% performance boost for German

  13. Sec. 2.2.1 Katakana Hiragana Kanji Romaji Language issues in tokenization • Chinese and Japanese have no spaces between words: • 莎拉波娃现在居住在美国东南部的佛罗里达。 • Not always guaranteed a unique tokenization • Further complicated in Japanese, with multiple alphabets intermingled フォーチュン500社は情報不足のため時間あた$500K(約6,000万円)

  14. Sec. 2.2.1 Language issues in tokenization • Arabic (or Hebrew) is basically written right to left, but with certain items like numbers written left to right • With modern Unicode representation concepts, the order of characters in files matches the conceptual order, and the reversal of displayed characters is handled by the rendering system, but this may not be true for documents in older encodings. • Other complexities that you know!

  15. Sec. 2.2.2 Stop words • In which stage? • With a stop list, you exclude from the dictionary entirely the commonest words. • Intuition: They have little semantic content: the, a, and, to, be • Using a stop list significantly reduces the number of postings that a system has to store, because there are a lot of them.

  16. Stop words • You need them for: • Phrase queries: “President of Iran” • Various song titles, etc.: “Let it be”, “To be or not to be” • “Relational” queries: “flights to London” • The general trend in IR systems: from large stop lists (200–300 terms) to very small stop lists (7–12 terms) to no stop list. • Good compression techniques (lecture 5) means the space for including stop words in a system is very small • Good query optimization techniques (lecture 7) mean you pay little at query time for including stop words.

  17. Sec. 2.2.3 Normalization to terms • In which stage? • We want to match U.S.A. and USA • Token normalization is the process of canonicalizingtokens so that matches occur despite superficial differences in the character sequences of the tokens • Result is terms: a term is a (normalized) word type, which is an entry in our IR system dictionary

  18. Normalization to terms • One way is using Equivalence Classes: • Searches for one term will retrieve documents that contain each of these members. • We most commonly implicitly define equivalence classes of terms rather than being fully calculated in advance (hand constructed), e.g., • deleting periods to form a term • U.S.A.,USA • deleting hyphens to form a term • anti-discriminatory, antidiscriminatory

  19. Sec. 2.2.3 Normalization: other languages • Accents: e.g., French résumé vs. resume. • Umlauts: e.g., German: Tuebingen vs. Tübingen • Normalization of things like date forms • 7月30日 vs. 7/30 • Tokenization and normalization may depend on the language and so is intertwined with language detection • Crucial: Need to “normalize” indexed text as well as query terms into the same form

  20. Sec. 2.2.3 Case folding • Reduce all letters to lower case • exception: upper case in mid-sentence? • Often best to lower case everything, since users will use lowercase regardless of ‘correct’ capitalization…

  21. Sec. 2.2.3 Normalization to terms • An alternative to equivalence classing is to do asymmetric expansion (hand constructed) • An example of where this may be useful • Enter: window Search: window, windows • Enter: windows Search: Windows, windows, window • Enter: Windows Search: Windows • Potentially more powerful, but less efficient

  22. Thesauri and soundex • Do we handle synonyms and homonyms? • E.g., by hand-constructed equivalence classes • car = automobile color = colour • What about spelling mistakes? • One approach is soundex, which forms equivalence classes of words based on phonetic heuristics • More in lectures 3 and 9

  23. Review • IR systems: • Indexing • Searching • Indexing: • Parsing document • Tokenization -> tokens • Normalization -> terms • Indexing -> index

  24. Review • Normalization: consider tokens rather than query. • Examples: • Case • Hyphen • Period • Synonyms • Spelling mistakes

  25. Review • Two methods for normalization: • Equivalence classing: often implicit. • Asymmetric expansion: • Query time: • a query expansion dictionary • more processing at query time • Indexing time: • more space for storing postings. • Asymmetric expansion is considerably less efficient than equivalence classing but more flexible.

  26. Stemming and lemmatization • Documents are going to use different forms of a word, such as organize, organizes, and organizing. • Additionally, there are families of derivationally related words with similar meanings, such as democracy, democratic, and democratization. • Reduce terms to their “roots” before indexing. • E.g., • am, are,is be • car, cars, car's, cars'car • the boy's cars are different colorsthe boy car be different color

  27. Sec. 2.2.4 Stemming • “Stemming” suggest crude affix chopping • language dependent • Example: • Porter’s algorithm • http://www.tartarus.org/~martin/PorterStemmer • Lovins stemmer • http://www.comp.lancs.ac.uk/computing/research/stemming/general/lovins.htm

  28. Sec. 2.2.4 Porter’s algorithm • Commonest algorithm for stemming English • Results suggest it’s at least as good as other stemming options • Conventions + 5 phases of reductions • phases applied sequentially • each phase consists of a set of commands • sample convention: Of the rules in a compound command, select the one that applies to the longest suffix.

  29. Sec. 2.2.4 Typical rules in Porter • sses  ss presses  press • ies  I bodies  bodi • ss  ss press  press • s  cats  cat • Many other rules are sensitive to the measure of words • (m>1) EMENT → • replacement → replac • cement → cement

  30. Sec. 2.2.4 Lemmatization • Reduce inflectional/variant forms to base form (lemma) properly with the use of a vocabulary and morphological analysis of words • Lemmatizer: a tool from Natural Language Processing which does full morphological analysis to accurately identify the lemma for each word.

  31. Stemming vs. Lemmatization • saw: • stemming might return just s, • Lemmatization would attempt to return either: • see: the use of the token was as a verb • saw: the use of the token was as a noun

  32. Helpfulness of normalization • Do stemming and other normalizations help? • Definitely useful for Spanish, German, Finnish, … • 30% performance gains for Finnish! • What about English?

  33. Helpfulness of normalization • English: • Not so considerable help! • Helps a lot for some queries, hurts performance a lot for others. • Stemming helps recall but harms precision • operative (dentistry) ⇒ oper • operational (research) ⇒ oper • operating (systems) ⇒ oper • For a case like this, moving to using a lemmatizer would not completely fix the problem

  34. Exercise 1 • Find rules for normalizing Farsi documents. • Please don’t copy! • Deadline: 21th Mordad, 11:00 AM

  35. Exercise 2 • Are the following statements true or false? Why? • a. In a Boolean retrieval system, stemming never lowers precision. • b. In a Boolean retrieval system, stemming never lowers recall. • c. Stemming increases the size of the vocabulary. • d. Stemming should be invoked at indexing time but not while processing a query

  36. Sec. 2.2.4 Language-specificity • Many of the above features embody transformations that are • Language-specific and • Often, application-specific • These are “plug-in” addenda to the indexing process • Both open source and commercial plug-ins are available for handling these

  37. Sec. 2.2 Dictionary entries – first cut These may be grouped by language (or not…). More on this in ranking/query processing.

  38. Faster postings merges:Skip lists

  39. Sec. 2.3 Recall basic merge • Walk through the two postings simultaneously, in time linear in the total number of postings entries 2 4 8 41 48 64 128 Brutus 2 8 1 2 3 8 11 17 21 31 Caesar If the list lengths are m and n, the merge takes O(m+n) operations. Can we do better? Yes

  40. Sec. 2.3 128 17 2 4 8 41 48 64 1 2 3 8 11 21 Augment postings with skip pointers (at indexing time) • Why? • To skip postings that will not figure in the search results. • How? • Where do we place skip pointers? • The resulted list is skip list. 128 41 31 11 31

  41. Sec. 2.3 17 2 4 8 41 48 64 1 2 3 8 11 21 But the skip successor of 11 on the lower list is 31, so we can skip ahead past the intervening postings. Query processing with skip pointers 128 41 128 31 11 31 Suppose we’ve stepped through the lists until we process 8 on each list. We match it and advance. We then have 41 and 11 on the lower. 11 is smaller.

  42. Sec. 2.3 Where do we place skips? • Tradeoff: • More skips  shorter skip spans  more likely to skip. But lots of comparisons to skip pointers. • Fewer skips  few pointer comparison, but then long skip spans  few successful skips.

  43. Sec. 2.3 Placing skips • Simple heuristic: for postings of length L, use L evenly-spaced skip pointers. • This ignores the distribution of query terms. • Easy if the index is relatively static; harder if L keeps changing because of updates. • This definitely used to help; with modern hardware it may not (Bahle et al. 2002) • The I/O cost of loading a bigger postings list can outweigh the gains from quicker in memory merging! D. Bahle, H. Williams, and J. Zobel. Efficient phrase querying with an auxiliary index. SIGIR 2002, pp. 215-221.

  44. Exercise • Do exercises 2.5 and 2.6 of your book. • Deadline: 23th Mordad, 11:00 AM

  45. Phrase queries and positional indexes

  46. Sec. 2.4 Phrase queries • Want to be able to answer queries such as “stanford university” – as a phrase • Thus the sentence “I went to university at Stanford” is not a match. • Most recent search engines support a double quotes syntax

  47. Phrase queries • PHRASE QUERIES has proven to be very easily understood and successfully used by users. • As many as 10% of web queries are phrase queries, • Many more queries are implicit phrase queries • For this, it no longer suffices to store only <term : docs> entries • Solutions?

  48. Sec. 2.4.1 A first attempt: Biword indexes • Index every consecutive pair of terms in the text as a phrase • For example the text “Friends, Romans, Countrymen” would generate the biwords • friends romans • romans countrymen • Each of these biwords is now a dictionary term • Two-word phrase query-processing is now immediate.

  49. Sec. 2.4.1 Longer phrase queries • The query “stanford university palo alto” can be broken into the Boolean query on biwords: stanford university AND university palo AND palo alto • Work fairly well in practice, • But there can and will be occasional false positives.

  50. Extended biwords • Now consider phrases such as “term in the vocabulary” • Perform part-of-speech-tagging (POST). • POST classify words as nouns, verbs, etc. • Group the terms into (say) Nouns (N) and articles/prepositions (X).

More Related