Agenda

29
Systems with Findability Measurement Shariq Bashir PhD-Student Technology University of Vienna

description

Evaluating Retrieval Systems with Findability Measurement Shariq Bashir PhD-Student Technology University of Vienna. Agenda. Document Findability Calculating Findability Measure GINI Coefficient Queries Creation for Findability Measure Experiments. Document Findability. - PowerPoint PPT Presentation

Transcript of Agenda

Page 1: Agenda

Evaluating Retrieval Systems

with Findability Measurement

Shariq BashirPhD-StudentTechnology University of Vienna

Page 2: Agenda

Agenda

• Document FindabilityDocument Findability• Calculating Findability Calculating Findability

MeasureMeasure• GINI CoefficientGINI Coefficient• Queries Creation for Findability Queries Creation for Findability

MeasureMeasure• ExperimentsExperiments

Page 3: Agenda

Document Findability

• Large Findability Large Findability of each and every document in Collection is considered an important factor in Legal or Patent Legal or Patent RetrievalRetrieval Settings.

• For example, in Patent Retrieval Settings, un-accessibility of a single related Patent document can approve wrong Patent application.

Page 4: Agenda

Document Findability

• Easy vs. Hard Findability– A patent is called easy Findablecalled easy Findable, if it is

accessible on top rank resultstop rank results of its several relevant queries.

– More the Patent will far away from the top rank results, the harder will be its Findabilityharder will be its Findability.

– Why?, because users are more interested on only top rank results (say top 30).

Page 5: Agenda

Document Findability

• Considered two Retrieval Systems (RS1, RS2) and three Patents (P1, P2, P3).

• The following table shows the Findability values of three Patents on top 30 results.

• It is clear, RS2 makes all Patents more Findable than RS1.

P1 P2 P3

RS1 0 2 7

RS2 4 7 5

Page 6: Agenda

What Makes Hard to Find Documents• System Bias

– Bias is a term used in IR, when retrieval system give preference to some featurespreference to some features of documents when it rank results of queries.

– Example, PageRankPageRank is bias toward larger in-links, BM25, BM25F, TF-IDFBM25, BM25F, TF-IDF are bias toward large terms frequencies.

– Bias is dangerous, why?why?, since under Bias some documents will be more findable, while rest of others will be very hard to find.

Page 7: Agenda

Bias with Findability analysis

• We can capture the bias capture the bias impact of different retrieval systems using Findability analysisFindability analysis.

• If a system has less biassystem has less bias, then it will make the individual documents more more FindableFindable.

• Findability evaluation vs. Precision based Findability evaluation vs. Precision based EvaluationEvaluation– We can’t use Findability evaluation at individual

queries level. – It is just large scale evaluation, only use for

capturing the bias of retrieval systems.

Page 8: Agenda

Findability Measure

• Given a collection of documents dD, with large set of Queries Q.

• kdq is the rank of dD in the result set of query qQ, c denotes the maximum rank that a user is willing to proceed down. The function f(kdq,c) returns a value of 1 if kdq<= c, and 0 otherwise.

Page 9: Agenda

GINI Coefficient

• For viewing the Bias of Retrieval System in a single value, we can use GINI coefficient.

• For computing GINI index, r(di) should be sort in ascending order. N total number of documents.

• If G = 0, then no bias, because all document are equally Findable. If G = 1, then only one document is Findable, and all other document have r(d) = 0.

Page 10: Agenda

Bias with Findability (Example)

r(d) with RS1

r(d) with RS2

d1 2 9

d2 0 7

d3 6 12

d4 5 14

D5 34 18

D6 4 11

d7 39 19

GINI 0.58 0.18

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

1 2 3 4 5 6 7

Document ordered by r(d)

Cu

mm

ula

tiv

e N

orm

ali

zed

r(d

)

Bias of RS1Bias of RS2No Bias

GINI Coefficient with Lorenz Curve

Page 11: Agenda

Bias of Retrieval Systems• Experiment Setting

– We used total Patents listed under United State Patent Classification (USPC) classes

– 433 (Dentistry), 424 (Drug, Bio-affecting and body treating compositions), 422 (Chemical apparatus and process disinfecting, deodorizing, preserving, or sterilizing), and 423 (Chemistry of inorganic compounds).

Page 12: Agenda

Experiment Setting

• Retrieval Systems used:– The OKAPI retrieval function (BM25BM25).– ExactExact match model.– TFIDFTFIDF– Language Modeling with term smoothing for Pseudo Relevance

Feedback selection (LMLM).– Kullback-Leibler divergence (KLDKLD).– Term selection value (Robertson and Walker) (QE TSQE TS).– Pseudo Relevance Feedback documents selection using

clustering approach (Cluster).

– For all Query Expansion models, we used top 35 top 35 documents for Pseudo relevance feedback and 50 terms 50 terms for query expansion.

Page 13: Agenda

Experiment Setting

• Queries Creation for Findability analysis

• In query creation, we try to reflect the approach of Patent Examiners, how they create their query sets during “Patent Invalidity Search”.

Page 14: Agenda

Experiment Setting

• Approach 1:– First, we extract all the single frequent terms single frequent terms from the Claim Claim

sections sections which have support greater than some threshold.– Then we combine these single frequent terms with two, three, and

four terms combinations for construction longer queries.

Patent (A)---------------------- Use Patent (A) as a

query for searching related documents.

Page 15: Agenda

Experiment Setting

Terms with Support >= 3

Page 16: Agenda
Page 17: Agenda

Experiment Setting

• Approach 2:– If patent contain many rare terms,

– then

– using queries collected from only single Patent document, we can’t search all of its similar Patentswe can’t search all of its similar Patents.

– In this Query Creation approach, we construct queries with considering Patent relatednessPatent relatedness.

Page 18: Agenda

Experiment Setting

• Approach 2 Steps:– (Step 1): For each Patent, group all of its related Patents in related Patents in

set (R) set (R) using k-nearest neighbor approach . – (Step 2): Then using this RR, construct its language model, for

finding dominant terms which can search the documents in R.

– Where Pjm(t|R) is the probability of term t in set R, and Pjm (t|corpos) is the probability of term t in whole collection.

– This is similar approach, as terms in Language Modeling (Query Expansion) are used for brining up relevant documents.

– (Step 3): Combine single terms with two, three, and four terms combinations for constructing longer queries.

Page 19: Agenda

Experiment Setting

• Properties of Queries used in Experiments

CQG 1: Approach 1CQG 2: Approach 2

Page 20: Agenda

Bias of Retrieval Systems with Patent Collection (433, 424)

With Query Creation Approach 1

Page 21: Agenda

Bias of Retrieval Systems with Patent Collection (433, 424)

With Query Creation Approach 1

Page 22: Agenda

Bias of Retrieval Systems with Patent Collection (433, 424)

With Query Creation Approach 1

Page 23: Agenda

Bias of Retrieval Systems with Patent Collection (433, 424)

With Query Creation Approach 2

Page 24: Agenda

Bias of Retrieval Systems with Patent Collection (433, 424)

With Query Creation Approach 2

Page 25: Agenda

Bias of Retrieval Systems with Patent Collection (433, 424)

With Query Creation Approach 2

Page 26: Agenda

GINI Index of Retrieval Systems with Patent Collection (433, 424)

Page 27: Agenda

GINI Index of Retrieval Systems with Patent Collection (422, 423)

Page 28: Agenda

Future Work

• We are working toward improving Findability of Patents using Query Expansion approach.

• We have results, in which selecting better documents for Pseudo Relevance Feedback can improve the Findability of documents.

• Considering external provided Ontology in Query Expansion, can also create its role in improving Findability of documents.

Page 29: Agenda

References

• Leif Azzopardi, Vishwa Vinay, Retrievability: an evaluation measure for higher order information access tasks, CIKM '08: Proceeding of the 17th ACM conference on Information and knowledge management, pages 561--570, October 26-30, 2008, Napa Valley, California, USA.

• Chris Jordan, Carolyn Wattters, Qigang Gao, Using controlled query generation to evaluate blind relevance feedback algorithms, JCDL '06: Proceedings of the 6th ACM/IEEE-CS joint conference on Digital libraries, 2006, Pages 286--295, Chapel Hill, NC, USA.

• Tonya Custis, Khalid Al-Kofahi, A new approach for evaluating query expansion: query-document term mismatch, SIGIR 2007: Proceedings of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 575--582, July 23-27, 2007, Amsterdam, The Netherlands.