Information Retrieval Introduction

Download Report

Transcript Information Retrieval Introduction

Instructor : Marina Gavrilova

     

Outline

Information Retrieval  IR Vs DBMS   Boolean Text Search Text Indexes  Simple relational text index  Example of inverted file Computing Relevance , The Vector Space Model Vector Space Model  Vector space documents and queries  Assigning weight to terms   TF xIDF Normalization Pair-wise documents similarity Text Clustering Probabilistic Models and Ranking Principles Iterative query refinement - Racchio Model  Query Modification  Collaborative Filtering and Ringo Collaborative Filtering

Goal

 Goal of this lecture is to introduce you to Informational retrieval and how it differentiates from DBMS. Then we will discuss how vector space model and Text Clustering help in computing relevance and similarity between documents. In the end, we will discuss how can queries be modified to provide end user with results that are similar to their interest.

Information Retrieval

  A research field traditionally separate from Databases  Goes back to IBM, Rand and Lockheed in the 50’s   G. Salton at Cornell in the 60’s Lots of research since then Products traditionally separate   Originally, document management systems for libraries, government, law, etc.

Gained prominence in recent years due to web search

IR vs. DBMS

 Seem like very different beasts: IR Imprecise Semantics Keyword search Unstructured data format Read-Mostly. Add docs occasionally Page through

top k

results DBMS Precise Semantics SQL Structured data Expect reasonable number of updates Generate full answer  Both support queries over large datasets, use indexing.  In practice, you currently have to choose between the two.

IR’s “Bag of Words” Model

   Typical IR data model:  Each document is just a bag (multiset) of words (“terms”) Detail 1: “Stop Words”    Certain words are considered irrelevant and not placed in the bag e.g., “the” e.g., HTML tags like

Detail 2: “Stemming” and other content analysis   Using English-specific rules, convert words to their basic form e.g., “surfing”, “surfed” --> “surf”

Boolean Text Search

 Find all documents that match a Boolean containment expression: “Windows” AND (“Glass” OR “Door”) AND NOT “Microsoft”  Note: Query terms are also filtered via stemming and stop words.

 When web search engines say “10,000 documents found”, that’s the Boolean search result size (subject to a common “max # returned’ cutoff).

Text “Indexes”

 When IR folks say “text index”…  Usually mean more than what DB people mean  In our terms, both “tables” and indexes  Really a logical schema (i.e., tables)  With a physical schema (i.e., indexes)  Usually not stored in a DBMS   Tables implemented as files in a file system We’ll talk more about this decision soon

A Simple Relational Text Index

    Create and populate a table InvertedFile(term string, docURL string ) Build a B+-tree or Hash index on InvertedFile.term

  Alternative 3 ( as entries in index) critical here for efficient storage!!  Fancy list compression possible, too Note: URL instead of RID, the web is your “heap file”!

 Can also cache pages and use RIDs This is often called an “inverted file” or “inverted index”  Maps from words -> docs Can now do single-word text search queries!

An Inverted File

 Search for  “databases”  “microsoft” term data database date day dbms decision demonstrate description design desire developer differ disability discussion division do document docURL http://www-inst.eecs.berkeley.edu/~cs186 http://www-inst.eecs.berkeley.edu/~cs186 http://www-inst.eecs.berkeley.edu/~cs186 http://www-inst.eecs.berkeley.edu/~cs186 http://www-inst.eecs.berkeley.edu/~cs186 http://www-inst.eecs.berkeley.edu/~cs186 http://www-inst.eecs.berkeley.edu/~cs186 http://www-inst.eecs.berkeley.edu/~cs186 http://www-inst.eecs.berkeley.edu/~cs186 http://www-inst.eecs.berkeley.edu/~cs186 http://www.microsoft.com

http://www-inst.eecs.berkeley.edu/~cs186 http://www.microsoft.com

http://www-inst.eecs.berkeley.edu/~cs186 http://www-inst.eecs.berkeley.edu/~cs186 http://www-inst.eecs.berkeley.edu/~cs186 http://www-inst.eecs.berkeley.edu/~cs186

Computing Relevance, Similarity: The Vector Space Model

Document Vectors

 Documents are represented as “bags of words”  Represented as vectors when used computationally  A vector is like an array of floating point  Has direction and magnitude  Each vector holds a place for every term in the collection  Therefore, most vectors are sparse

Document Vectors: One location for each word.

A B C D I E F G H

nova 10 5 5 galaxy heat 5 3 10 h’wood film role 6 10 8 7 “Nova” occurs 10 times in text A “Galaxy” occurs 5 times in text A “Heat” occurs 3 times in text A 9 10 2 7 8 5 diet 10 9 1 fur 10 10 3

Document Vectors

Document ids

A B C D I E F G H

nova 10 5 5 galaxy heat 5 3 10 6 7 10 h’wood film 10 9 2 7 8 10 8 5 role 7 5 9 diet fur 10 9 1 10 10 3

We Can Plot the Vectors

Star Doc about movie stars Doc about astronomy Doc about mammal behavior Diet Assumption: Documents that are “close” in space are similar.

Vector Space Model

 Documents are represented as

vectors

in term space   Terms are usually stems Documents represented by binary vectors of terms  Queries represented the same as documents  A vector distance measure between the query and documents is used to rank retrieved documents   Query and Document similarity is based on length and direction of their vectors Vector operations to capture boolean query conditions  Terms in a vector can be “weighted” in many ways

Vector Space Documents and Queries

docs

D1 D2 D3 D4 D5 D6 D7 D8 D9 D10 D11

Q t1

1 1 0 1 1 1 0 0 0 0 1 1

q1 t2

0 0 1 0 1 1 1 1 0 1 0 2

q2 t3

1 0 1 0 1 0 0 0 1 1 1 3

q3

RSV=Q.Di

4 1 5 1 6 3 2 2 3 5 3 t 3 D 9 D 11 D 3 D 10 D 7 D 1 D 5 D 8 D 6 D 2 D 4 t 2 t 1

Boolean term combinations Q is a query – also represented as a vector

Assigning Weights to Terms

 Binary Weights  Raw term frequency  tf x idf  Recall the Zipf distribution  Want to weight terms highly if they are  frequent in relevant documents … BUT  infrequent in the collection as a whole

Binary Weights

 Only the presence (1) or absence (0) of a term is included in the vector

docs

D1 D2 D3 D4 D5 D6 D7 D8 D9 D10 D11 0 0 0 0 1

t1

1 1 0 1 1 1 1 1 0 1 0

t2

0 0 1 0 1 1 0 0 1 1 1

t3

1 0 1 0 1 0

Raw Term Weights

 The frequency of occurrence for the term in each document is included in the vector

docs

D1 D2 D3 D4 D5 D6 D7 D8 D9 D10 D11 0 0 0 0 4

t1

2 1 0 3 1 3 8 10 0 3 0

t2

0 0 4 0 6 5 0 0 1 5 1

t3

3 0 7 0 3 0

TF x IDF Normalization

 Normalize the term weights (so longer documents are not unfairly given more weight)  The longer the document, the more likely it is for a given term to appear in it, and the more often a given term is likely to appear in it. So, we want to reduce the importance attached to a term appearing in a document based on the length of the document.

w ik

tf ik

t k

 1

(

tf ik

log(

N

/ )

2

[log(

n k N

) /

n k

)]

2

Pair-wise Document Similarity

A B C D

nova 1 5 galaxy 3 2 heat 1 h’wood film role 2 4 1 1 5 diet fur

How to compute document similarity?

Pair-wise Document Similarity

D D

2 1 

w

11 ,

w

12 , ...,

w

1

t

w

21 ,

w

22 , ...,

w

2

t sim

(

D

1 ,

D

2 ) 

i t

  1

w

1

i

w

2

i sim

(

A

,

B

)  ( 1  5 )  ( 2  3 )  11

sim

(

A

,

C

)  0

sim

(

A

,

D

)  0

sim

(

B

,

C

)  0

sim

(

B

,

D

)  0

sim

(

C

,

D

)  ( 2  4 )  ( 1  1 )  9

A B C D

nova 1 5 galaxy heat 3 1 2 h’wood film role 2 4 1 1 5 diet fur

Pair-wise Document Similarity (cosine normalization)

D D

2 1 

w

11 ,

w

12 , ...,

w

1

t

w

21 ,

w

22 , ...,

w

2

t sim

(

D

1 ,

D

2 ) 

i t

  1

w

1

i

w

2

i

unnormaliz

sim

(

D

1 ,

D

2 )  ed

i t

  1

i t

  1

w

1

i

w

2

i

(

w

1

i

) 2 

i t

  1 (

w

2

i

) 2 cosine normalized

Text Clustering

 Finds overall similarities among groups of documents  Finds overall similarities among groups of tokens  Picks out some themes, ignores others

Text Clustering

Clustering is “The art of finding groups in data.” -- Kaufmann and Rousseeu

Term 1 Term 2

Problems with Vector Space

 There is no real theoretical basis for the assumption of a term space  It is more for visualization than having any real basis  Most similarity measures work about the same  Terms are not really orthogonal dimensions  Terms are not independent of all other terms; remember our discussion of correlated terms in text

Probabilistic Models

 Rigorous formal model attempts to predict the probability that a given document will be relevant to a given query  Ranks retrieved documents according to this probability of relevance (Probability Ranking Principle)  Relies on accurate estimates of probabilities

Probability Ranking Principle

 If a reference retrieval system’s response to each request is a ranking of the documents in the collections in the order of decreasing probability of usefulness to the user who submitted the request, where the probabilities are estimated as accurately as possible on the basis of whatever data has been made available to the system for this purpose, then the overall effectiveness of the system to its users will be the best that is obtainable on the basis of that data.

Stephen E. Robertson,

J. Documentation

1977

Query Modification

 Problem: How can we reformulate the query to help a user who is trying several searches to get at the same information?

 Thesaurus expansion:   Suggest terms similar to query terms Relevance feedback:  Suggest terms (and documents) similar to retrieved documents that have been judged to be relevant

Relevance Feedback

 Main Idea:  Modify existing query based on relevance judgements   Extract terms from relevant documents and add them to the query AND/OR re-weight the terms already in the query  There are many variations:   Usually positive weights for terms from relevant docs  Sometimes negative weights relevant docs for terms from non Users, or the system, guide this process by selecting terms from an automatically-generated list.

Rocchio Method

 Rocchio automatically  Re-weights terms  Adds in new terms (from relevant docs)  have to be careful when using negative terms  Rocchio is not a machine learning algorithm

Rocchio Method

Q

1  

Q

0  

n

1

n

1

i

  1

R i

 

n

2

i n

 2  1

S i where Q

0  the vector for the initial query

R i

 the vector for the relevant document

i S i

 the vector for the non relevant document

i n

1  the number of relevant documents chosen

n

2  ,  the number  and  of non relevant tune the importance documents chosen of relevant and nonrelevan (in some studies best to set  to 0.75

and  to 0.25) t terms

Alternative Notions of Relevance Feedback

 Find people whose taste is “similar” to yours.

 Will you like what they like?

 Follow a user’s actions in the background.  Can this be used to predict what the user will want to see next?

 Track what lots of people are doing.  Does this implicitly indicate what they think is good and not good?

Collaborative Filtering (Social Filtering)

 If Pam liked the paper, I’ll like the paper  If you liked Star Wars, you’ll like Independence Day  Rating based on ratings of similar people  Ignores text, so also works on sound, pictures etc.

 But: Initial users can bias ratings of future users

Star Wars Jurassic Park Terminator II Independence Day 6 3 7 Sally 7 4 4 7 Bob 7 7 7 2 Chris 3 4 6 2 Lynn 4 4 3 ?

Karen 7

Ringo Collaborative Filtering

Users rate items from like to dislike  7 = like; 4 = ambivalent; 1 = dislike    A normal distribution; the extremes are what matter Nearest Neighbors Strategy: Find similar users and predicted (weighted) average of user ratings Pearson Algorithm: U and user J Weight by degree of correlation between user  1 means similar, 0 means no correlation, -1 dissimilar  Works better to compare against the ambivalent rating (4), rather than the individual’s average score

r UJ

   (

U

(

U

U

 )

U

2  )( 

J

 (

J J

 )

J

) 2

Computing Relevance

  Relevance calculation involves how often search terms appear in doc, and how often they appear in collection:  More search terms found in doc  doc is more relevant  Greater importance attached to finding rare terms Doing this efficiently in current SQL engines is not easy:  “Relevance of a doc wrt a search term” is a function that is called once per doc the term appears in (docs found via inv. index):  For efficient fn computation, for each term, we can store the # times it appears in each doc, as well as the # docs it appears in.

   Must also sort retrieved docs by their relevance value.

Also, think about Boolean operators (if the search has multiple terms) and how they affect the relevance computation!

An object-relational or object-oriented DBMS with good support for function calls is better, but you still have long execution path-lengths compared to optimized search engines.

Updates and Text Search

  Text search engines are designed to be query-mostly:  Deletes and modifications are rare     Can postpone updates (nobody notices, no transactions!)  Updates done in batch (rebuild the index) Can’t afford to go off-line for an update?

  Create a 2nd index on a separate machine Replace the 1st index with the 2nd!

So no concurrency control problems Can compress to search-friendly, update-unfriendly format Main reason why text search engines and DBMSs are usually separate products.

 Also, text-search engines tune that one SQL query to death!

DBMS vs. Search Engine Architecture

DBMS Search Engine Query Optimization and Execution { Relational Operators Files and Access Methods Buffer Management Disk Space Management Concurrency and Recovery Needed Search String Modifier Ranking Algorithm “The Query” The Access Method Buffer Management OS } Simple DBMS Disk Space Management

IR vs. DBMS Revisited

  Semantic Guarantees  DBMS guarantees transactional semantics  If inserting Xact commits, a later query will see the update   Handles multiple concurrent updates correctly IR systems do not do this; nobody notices!

  Postpone insertions until convenient No model of correct concurrency Data Modeling & Query Complexity  DBMS supports any schema & queries   Requires you to define schema  Complex query language hard to learn IR supports only one schema & query   No schema design required (unstructured text) Trivial to learn query language

Lots More in IR …

      How to “rank” the output? I.e., how to compute relevance of each result item w.r.t. the query?

 Doing this well / efficiently is hard!

Other ways to help users paw through the output?

 Document “clustering”, document visualization How to take advantage of hyperlinks?

 Really cute tricks here!

How to use compression for better I/O performance?

 E.g., making RID lists smaller  Try to make things fit in RAM!

How to deal with synonyms, misspelling, abbreviations? How to write a good web crawler?

Summary

 First we studied difference between Information Retrieval and DBMS . Then we disused on two type of searches (Boolean and Text based search) used in DBMS .  In addition, we learned how we can compute relevance between documents based on words using Vector Space Model and how text clustering can be used to find similarity between documents and in the end we discussed Racchio Model for iterative query refinement.

 Thanks You !