1 / 33

Application of NLP in Information Retrieval

Application of NLP in Information Retrieval. Nirdesh Chauhan Ajay Garg Veeranna A.Y. Neelmani Singh. Presentation Outline. Overview of current IR Systems Problems with NLP in IR Major applications of NLP in IR. Motivation.

fern
Download Presentation

Application of NLP in Information Retrieval

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Application of NLP in Information Retrieval Nirdesh Chauhan Ajay Garg Veeranna A.Y. Neelmani Singh

  2. Presentation Outline • Overview of current IR Systems • Problems with NLP in IR • Major applications of NLP in IR

  3. Motivation • Most successful general purpose retrieval methods are statistical methods. • Sophisticated linguistic processing often degrade performance.

  4. What is IR ?? • “Information retrieval system is one that searches a collection of natural language documents with the goal of retrieving exactly the set of documents that pertain to a users question” • Have their origins in library systems • Do not attempt to deduce or generate answers

  5. Basics of IR Systems

  6. Basics of IR Systems (contd…) • Indexing the collection of documents. • Transforming the query in the same way as the document content is represented. • Comparing the description of each document with that of the query. • Listing the results in order of relevancy.

  7. Basics of IR Systems (contd…) • Retrieval Systems consist of mainly two processes: • Indexing • Matching

  8. Indexing • Indexing is the process of selecting terms to represent a text. • Indexing involves: • Tokenization of string • Removing frequent words • Stemming • Two common Indexing Techniques: • Boolean Model • Vector space model

  9. Information Retrieval Models • A retrieval model consists of: • D: representation for documents • R: representation for queries • F: a modeling framework for D, Q • R(q, di): a ranking or similarity function which orders the documents with respect to a query.

  10. Boolean Model • Queries are represented as Boolean combinations of the terms. • Set of documents that satisfied the Boolean expression are retrieved in response to the query. • Drawback • User is given no indication as to whether some documents in the retrieved set are likely to be better than others in the set

  11. Vector Space Model • In this model documents and queries are represented by vectors in T dimensional space. • T is the number of distinct terms used in the documents. • Each axis corresponds to one term. • Ranked list of documents ordered by similarity to the query where similarity between a query and a document is computed using a metric on the respective vectors.

  12. Matching • Matching is the process of computing a measure of similarity between two text representations. • Relevance of a document is computed based on following parameters: • tf - term frequency is simply the number of times a given term appears in that document.tfi.j = (count of ith term in jth document)/(total terms in jth document) • idf - inverse document frequency is a measure of the general importance of the termidfi = (total no. of documents)/(no. of documents containing ith term) • tfidfi,j score = tf * idf

  13. Evaluation of IR Systems • Two common effectiveness measures include: • Precision: Proportion of retrieved documents that are relevant. • Recall: Proportion of relevant documents that are retrieved. • Ideally both precision and recall should be 1. • In practice, these are inversely related.

  14. Problems regarding NLP in IR • Linguistic techniques must be essentially perfect • Errors occurs in linguistic processing e.g. POS tagging, sense resolution, parsing etc. • Effect of these errors on retrieval performance must be considered. • Incorrectly resolving two usages of the same sense differently is disastrous for retrieval effectiveness. • Disambiguation accuracy of at least 90% is required just to avoid degrading retrieval effectiveness.

  15. Problems regarding NLP in IR (contd…) • Queries are difficult • Queries are especially troublesome for most NLP processing. • They are generally quite short and offer little to assist linguistic processing. • But to have any effect whatsoever on retrieval queries must also contain the type of index terms used in documents. • Compensated by query expansion and blind feedback.

  16. Problems regarding NLP in IR (contd…) • Linguistic knowledge is implicitly exploited • Statistical techniques implicitly exploit the same information the linguistic techniques make explicit. • So linguistic techniques may provide little benefit over appropriate statistical techniques.

  17. Problems regarding NLP in IR (contd…) • Term normalization might be beneficial. • Map various formulations and spellings of a same lexical item to a common form. • E.g. somatotropin and somatotrophin analyzer and analyser

  18. Application of NLP in IR • We discuss here the following applications: • Conceptual Indexing • Enhancement in Matching • Semantically Relatable Sets

  19. Conceptual Indexing • Matching of concepts in document and query instead of matching words. • Use of WORDNET synsets as concepts. • Word Sense Disambiguation for nouns: noun disambiguated to a single synset.

  20. Conceptual Indexing • Extended Vector space model. • Query and Document represented as set of vectors, each of them representing different aspects of them. • stems of words not found in WordNet or not disambiguated. • synonym set ids of disambiguated nouns. • stems of the disambiguated nouns. • Weights are applied to similarity measure of corresponding vector. • Failed w.r.to stemming due to poor disambiguation

  21. Enhancement in Matching • For example, if index terms are noun phrases then a partial match may be made if two terms share a common head but are not identical.

  22. Semantically Relatable Sets • This method enhances indexing. • Documents and queries are represented as Semantically Relatable Sets (SRS). • Example “A new book on IR” • SRS corresponding to this query are:{A, book}, {new, book}, {book, on, IR}

  23. SRS Based Search The relevance score for a document d, where Rq(d) = Relevance of the document d to the query q |Sd| = Number of sentences in the document d rq(s) = Relevance of sentence s to the query q The relevance of the sentence s to the query q where weight(srs) = weight of the SRS srs depending on its type. press(srs) = 1 if srs is present in sentence s, 0 otherwise.

  24. Improving performance of SRS based Search • Stemming • Words in document and query SRS are stemmed based on WordNet. • Takes care of the morphological divergence problem. • “children_NN” stemmed to “child_NN”, but the word “childish_JJ” will not be stemmed to “child_NN”, since the word “childish” is an adjective, whereas “child” is a noun. • Using Word Similaritysynonymy/hypernymy/hyponymy problem is tackled by this method. The relevance of the sentence s to the query q is reformulated as: t() is the SRS similarity measure , t(srs,srs’) = (cw1,cw1’)*equal(fw,fw’)*t(cw2,cw2’) For (FW,CW) matching, t(cw1,cw1’) is set to one and for (CW,CW) matching, equal(fw,fw’) is set to one. In all other cases, t(w1,w2) gives the relatedness measure of w1 and w2 (calculated using the baseline similarity measure “path”).

  25. Improving performance of SRS based Search (contd…) • SRS Augmentation • Rule: (noun1, in/on, noun2) => (noun2, noun1) • Example: (defeat, in, election) will create an augmented SRS as (election, defeat) • Rule: (adjective, noun) => (noun, adjective_in_noun_form) • Example: (polluted, water) will augment (water, pollution) • Rule: (adjective, with, noun–(ANIMATE)) => (noun, adjective_in_noun_form) • Example: (angry, with, result) will augment (result, anger), whereas (angry, with, John) will not augment (John, anger).

  26. Case Study Query: I need to know the gas mileage for my audi a8 2004 model Source: Yahoo search (search.yahoo.com)

  27. Case Study (contd…) Query: I need to know the gas mileage for my audi a8 2004 model Source: Y!Q search (yq.search.yahoo.com)

  28. Case Study (contd…) Query: I need to know the gas mileage for my audi a8 2004 model Source: Google search (www.google.com)

  29. Case Study (contd…) • Yahoo Search • Pure text-based search. • Result generates instance of same text containing documents. • Y!Q Search • Use of semantics but not efficient. • Attempts to generate answer. However this is done less efficiently here. • Google Search • Efficient use of NLP for deduction of answer form given question. • A step towards question-answering !!

  30. Conclusion • Research efforts to address appropriate tasks are underway.E.g. document summarization, generating answers. • Achieving extremely efficient NLP techniques is an idealization.

  31. References • Voorhees, EM, "Natural Language Processing and Information Retrieval," in Pazienza, MT (ed.), Information Extraction: Towards Scalable, Adaptable Systems, New York: Springer, 1999. • Salton G Wong A Yang CS A Vector Space Model for Automatic Indexing Communications of the ACM (1975) 613-620. • Mari Vallez; Rafael Pedraza-Jimenez. Natural Language Processing in Textual Information Retrieval and Related Topics "Hipertext.net", num. 5, 2007. • Sanjeet Khaitan, Kamaljeet Verma and Pushpak Bhattacharyya, Exploiting Semantic Proximity for Information Retrieval, IJCAI 2007, Workshop on Cross Lingual Information Access, Hyderabad, India, Jan, 2007. • Wikipedia

  32. Questions ??

  33. Thank You !!!!!

More Related