CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

49
CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002

Transcript of CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Page 1: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

CS276AText Information Retrieval, Mining, and

Exploitation

Lecture 831 Oct 2002

Page 2: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Recap: IR based on Language Model

query

d1

d2

dn

Information need

document collection

generationgeneration

)|( dMQP 1dM

2dM

ndM One night in a hotel, I saw this late night talk show where Sergey Brin popped on suggesting the web search tip that you should think of some words that would likely appear on pages that would answer your question and use those as your search terms – let’s exploit that idea!

Page 3: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Recap: Query generation probability (unigrams)

Ranking formula

The probability of producing the query given the language model of document d using MLE is:

Qt d

dt

Qtdmld

dl

tf

MtpMQp

),(

)|(ˆ)|(ˆ

Unigram assumption:Given a particular language model, the query terms occur independently

),( dttf

ddl

: language model of document d

: raw tf of term t in document d

: total number of tokens in document d

dM

)|()(

)|()(),(

dMQpdp

dQpdpdQp

Page 4: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Recap: Query generation probability (mixture model)

P(w|d) = Pmle(w|Md) + (1 – )Pmle(w|Mc) Mixes the probability from the document with

the general collection frequency of the word. Correctly setting is very important A high value of lambda makes the search

“conjunctive-like” – suitable for short queries A low value is more suitable for long queries Can tune to optimize performance

Perhaps make it dependent on document size (cf. Dirichlet prior or Witten-Bell smoothing)

Page 5: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Today’s topics

Relevance Feedback Query Expansion

Page 6: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Relevance Feedback

Relevance feedback = user feedback on relevance of initial set of results

The user marks returned documents as relevant or non-relevant.

The system computes a better representation of the information need based on feedback.

Relevance feedback can go through one or more iterations.

Page 7: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Relevance Feedback: Example

Image search engine (couldn’t find relevance feedback engine for text!)

url: http://nayana.ece.ucsb.edu/imsearch/imsearch.html

Page 8: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Initial Query

Page 9: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Results for Initial Query

Page 10: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Relevance Feedback

Page 11: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Results after Relevance Feedback

Page 12: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Rocchio Algorithm

The Rocchio algorithm incorporates relevance feedback information into the vector space model.

The optimal query vector for separating relevant and non-relevant documents:

Unrealistic: we don’t know all relevant documents.

Page 13: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Rocchio Algorithm

Used in practice:

Typical weights: alpha = 8, beta = 64, gamma = 64 Tradeoff alpha vs beta/gamma: If we have a lot of

judged documents, we want a higher beta/gamma. But we usually don’t …

Page 14: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Relevance Feedback inProbabilistic Information Retrieval

How?

Page 15: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Relevance Feedback inProbabilistic Information Retrieval

We can modify the query based on relevance feedback and apply standard model.

Examples: Binary independence model Language model

Page 16: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Binary Independence Model

n

i i

i

qNRxp

qRxpqROdqRO

1 ),|(

),|()|(),|(

• Since xi is either 0 or 1:

01 ),|0(

),|0(

),|1(

),|1()|(),|(

ii x i

i

x i

i

qNRxp

qRxp

qNRxp

qRxpqROdqRO

Page 17: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Binary Independence Model

01 ),|0(

),|0(

),|1(

),|1()|(),|(

ii x i

i

x i

i

qNRxp

qRxp

qNRxp

qRxpqROdqRO

Used as before. We assume this is =1in simple retrieval.In relevancefeedback, wehave data to estimateprobability ofnon-occurrence.

Page 18: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Binary Independence Model

ii q i

i

q i

i

qNRxp

qNRxp

qRxp

qRxpqROdqRO

),|1(

),|0(

),|0(

),|1()|(),|(

Note that we have 3 relevance “states” now: relevant, non-relevant and unjudged.

Page 19: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Positive vs Negative Feedback

Positive feedback is more valuable than negative feedback.

Many systems only allow positive feedback.

Why?

Page 20: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Relevance Feedback: Assumptions

A1: User has sufficient knowledge for initial query.

A2: Relevance prototypes are “well-behaved”. Either: All relevant documents are similar to a

single prototype. Or: There are different prototypes, but they

have significant vocabulary overlap.

Page 21: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Violation of A1

User does not have sufficient initial knowledge.

Examples: Misspellings (Brittany Speers) Cross-language information retrieval (hígado) Mismatch of searcher’s vocabulary vs

collection vocabulary (e.g., regional differences, different fields of scientific study: genetics vs medicine, bernoulli naïve bayes vs binary independence model)

Page 22: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Violation of A2

There are several relevance prototypes. Examples:

Burma/Myanmar Contradictory government policies Pop stars that worked at Burger King

Often: instances of a general concept Good editorial content can address problem

Report on contradictory government policies

Page 23: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Relevance Feedback on the Web

Some search engines offer a similar/related pages feature (simplest form of relevance feedback)

Google (link-based) Altavista Stanford web

But some don’t because it’s hard to explain to average user:

Alltheweb msn Yahoo

Excite initially had true relevance feedback, but abandoned it due to lack of use.

Page 24: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Relevance Feedback: Cost

Long queries are inefficient for typical IR engine.

Long response times for user. High cost for retrieval system.

Why?

Page 25: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Other Uses of Relevance Feedback

Following a changing information need Maintaining an information filter (e.g., for a

news feed) Active learning Topics for next quarter

Page 26: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Active Learning

Goal: create a training set for a text classifier (or some other classification problem)

One approach: uncertainty sampling At any point in learning, present document with

highest uncertainty to user and get relevant/non-relevant judgment. (closest to p( R) = 0.5 )

Scarce resource is user’s time: maximize benefit of each decision.

Active learning significantly reduces the number of labeled documents needed to learn a category.

Page 27: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Active Learning

documents

classifier selects most uncertain

document

most uncertaindocument

retrain classifier

user labelsdocument

somehow BuildInitial classifier

Page 28: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Active Learning

Could we use active learning for relevance feedback?

Page 29: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Relevance FeedbackSummary

Relevance feedback has been shown to be effective at improving relevance of results.

Full relevance feedback is painful for the user.

Full relevance feedback is not very efficient in most IR systems.

Other types of interactive retrieval may improve relevance by as much with less work.

Page 30: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Forward Pointer: DirectHit

DirectHit uses indirect relevance feedback. DirectHit ranks documents higher that

users look at more often. Not user or query specific.

Page 31: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Pseudo Feedback

Pseudo feedback attempts to automate the manual part of relevance feedback.

Retrieve an initial set of relevant documents. Assume that top-ranked documents are

relevant.

Page 32: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Pseudo Feedback

documents

retrievedocuments

highest rankeddocuments

apply relevance feedback

label top kdocs relevant

initialquery

Page 33: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Pseudo Feedback

Not surprisingly, the success of pseudo feedback depends on whether relevance assumption is true.

If it is true, pseudo feedback can improve precision and recall dramatically.

If it is not true, then the results will become less relevant in each iteration. (concept drift)

Example: ambiguous words (“jaguar”) Bimodal distribution: depending on query,

performance is dramatically better or dramatically worse.

Unfortunately, hard to predict which will be the case.

Page 34: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Pseudo-Feedback: Performance

Page 35: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Query Expansion

In relevance feedback, users give additional input (relevant/non-relevant) on documents.

In query expansion, users give additional input (good/bad search term) on words or phrases.

Page 36: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Query Expansion: Example

Also: see altavista, teoma

Page 37: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Types of Query Expansion

Refinements based on query log mining Common on the web

Global Analysis: Thesaurus-based Controlled vocabulary

Maintained by editors (e.g., medline) Automatically derived thesaurus

(co-occurrence statistics)

Local Analysis: Analysis of documents in result set

Page 38: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Controlled Vocabulary

Page 39: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Automatic Thesaurus Generation

High cost of manually producing a thesaurus Attempt to generate a thesaurus automatically

by analyzing a collection of documents Two main approaches

Co-occurrence based (co-occurring words are more likely to be similar)

Shallow analysis of grammatical relations Entities that are grown, cooked, eaten, and digested

are more likely to be food items.

Co-occurrence based is more robust, grammatical relations are more accurate.

Why?

Page 40: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Co-occurrence Thesaurus

Simplest way to compute one is based on term-term similarities in C = AAT where A is term-document matrix.

wi,j = (normalized) weighted count (ti , dj)

ti

djn

m

Page 41: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Automatic Thesaurus GenerationExample

Page 42: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Automatic Thesaurus GenerationDiscussion

Quality of associations is usually a problem. Very similar to LSI (why?) Problems:

“False positives” Words deemed similar that are not

“False negatives” Words deemed dissimilar that are similar

Page 43: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Query Expansion: Summary

Query expansion is very effective in increasing recall.

In most cases, precision is decreased, often significantly.

Page 44: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Sense-Based Retrieval

In query expansion, new words are added to the query (disjunctively). Increase of matches.

In sense-based retrieval, term matches are only counted if the same sense is used in query and document. Decrease of matches.

Example: In sense-based retrieval, “jaguar” is only a match if it’s used in the “animal” sense in both query and document.

Page 45: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Sense-Based Retrieval: Results

Page 46: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Expansion vs. Sense-Based Retrieval

Same type of information is used in pseudo relevance feedback and sense-based retrieval.

But: disambiguation is expensive Indexing with senses is complicated Automatic sense-based retrieval only makes

sense for long queries If senses are supplied in interactive loop,

then it’s easier to add words rather than senses

Why?

Page 47: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Exercise

What is the factor of increase of the index size?

How could you integrate confidence numbers for senses?

Page 48: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Interactive Retrieval

Query expansion and relevance feedback are examples of interactive retrieval.

Others: Query “editing” (adding and removing terms) Visualization-based, e.g., interaction with

visual map Parametric search

Page 49: CS276A Text Information Retrieval, Mining, and Exploitation Lecture 8 31 Oct 2002.

Resources

MG Ch. 4.7MIR Ch. 5.2 – 5.4Singhal, Mitra, Buckley: Learning routing queries in a query

zone, ACM SIGIR, 1997.Yonggang Qiu , Hans-Peter Frei, Concept based query

expansion, Sigir, 1993.Schuetze: Automatic Word Sense Discrimination,

Computational Linguistics, 1998.Buckley, Singhal, Mitra, Salton, New retrieval approaches

using smart: trec4, nist, 1996.Gerard Salton and Chris Buckley. Improving retrieval

performance by relevance feedback. Journal of the American Society ]or In:formation Science, 41(4):288-297, 1990.