Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and...
-
date post
19-Dec-2015 -
Category
Documents
-
view
216 -
download
3
Transcript of Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and...
![Page 1: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/1.jpg)
Chapter 8
Evaluating Search Engine
![Page 2: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/2.jpg)
2
Evaluation Evaluation is key to building effective and efficient
search engines
Measurement usually carried out in controlled laboratory experiments
Online testing can also be done
Effectiveness, efficiency, and cost are related
e.g., if we want a particular level of effectiveness and efficiency, this will determine the cost of the
system configuration
Efficiency and cost targets may impact effectiveness
![Page 3: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/3.jpg)
3
Evaluation Corpus Test collections consisting of documents, queries, and
relevance judgments, e.g.,
![Page 4: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/4.jpg)
4
Test Collections
![Page 5: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/5.jpg)
5
TREC Topic Example
![Page 6: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/6.jpg)
6
Relevance Judgments Obtaining relevance judgments is an expensive, time-
consuming process Who does it?
What are the instructions?
What is the level of agreement?
TREC judgments Depend on task being evaluated, e.g., topical relevance
Generally binary, i.e., relevant vs. non-relevant
Agreement good because of “narrative”
![Page 7: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/7.jpg)
7
Pooling Exhaustive judgments for all documents in a collection
is not practical
Pooling technique is used in TREC
Top k results (for TREC, k varied between 50 and 200) from the rankings obtained by different search engines (or retrieval algorithms) are merged into a pool
Duplicates are removed
Documents are presented in some random order to the relevance judges
Produces a large number of relevance judgments for each query, although still incomplete
![Page 8: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/8.jpg)
8
Query Logs Used for both tuning and evaluating search engines
Also for various techniques such as query suggestion
Typical contents
User identifier or user session identifier
Query terms - stored exactly as user entered
List of URLs of results, their ranks on the result list and whether they were clicked on
Timestamp(s) - records the time of user events such as query submission, clicks
![Page 9: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/9.jpg)
9
Query Logs Clicks are not relevance judgments
Although they are correlated
Biased by a number of factors, such as rank on result list
Can use clickthough data to predict preferences between pairs of documents
Appropriate for tasks with multiple levels of relevance, i.e., closely-relevant, relevant, semi-relevant, non-
relevant, etc. focused on user relevance
Various “policies” used to generate preferences
![Page 10: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/10.jpg)
10
Example Click Policy Skip Above and Skip Next
Click data
Generated preferences
![Page 11: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/11.jpg)
11
Effectiveness MeasuresA is set of relevant documents, B is set of retrieved documents
Relevant Non-Relevant
Retrieved A B A BNot Retrieved A B A B
![Page 12: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/12.jpg)
12
Classification Errors False Positive (Type I Error)
Non-relevant documents retrieved
False Negative (Type II Error)
Relevant documents not retrieved
1 - Recall
Precision is used when probability that a positive result is correct is important
![Page 13: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/13.jpg)
13
F Measure Harmonic mean of recall and precision
Harmonic mean emphasizes the importance of small values, whereas the arithmetic mean is
affected more by outliers that are unusually large
More general form
β is a parameter that determines relative importance of recall and precision
What if β = 1?
![Page 14: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/14.jpg)
14
Ranking Effectiveness
![Page 15: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/15.jpg)
15
Summarizing a Ranking
Calculating recall and precision at fixed rank positions
Calculating precision at standard recall levels, from 0.0 to 1.0
Requires interpolation
Averaging the precision values from the rank positions where a relevant document was retrieved
![Page 16: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/16.jpg)
16
Average Precision
![Page 17: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/17.jpg)
17
Averaging Across Queries
![Page 18: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/18.jpg)
18
Averaging Mean Average Precision (MAP)
Summarize rankings from multiple queries by averaging average precision
Most commonly used measure in research papers
Assumes user is interested in finding many relevant documents for each query
Requires many relevance judgments in text collection
Recall-precision graphs are also useful summaries
![Page 19: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/19.jpg)
19
MAP
![Page 20: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/20.jpg)
20
Recall-Precision Graph Too much information could be lost using MAP
Recall-precision graphs provide more detail on the effectiveness of the ranking at different recall levels
Example. The recall-precision graphs for the two previous queries
Ranking of Query 1
Ranking of Query 2
![Page 21: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/21.jpg)
21
Interpolation To average graphs, calculate precision at standard recall
levels (from 0.0 to 1.0 with an increments of 0.1):
where R is a recall level and S is the set of observed (R’, P’) points
Defines precision at any recall level as the maximum precision observed in any recall-precision
point at a higher recall level
• Produces a step function
• Defines precision as 1.0 at recall level 0.0, unless stated otherwise
![Page 22: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/22.jpg)
22
Interpolation Example. Interpolated recall-precision graphs for the
two previous queries
![Page 23: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/23.jpg)
23
Average Precision @ Standard Recall Levels The interpolation yields the best possible precision
values at different recall levels
The average precision values at the standard recall levels are calculated by averaging the
precision values for each query
InterpolatedPrecisionvalues
Average Precision values
![Page 24: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/24.jpg)
24
Average Recall-Precision Graph Example. Recall-precision graph plotted by simply joining the
average precision points at the standard recall levels
0.55
![Page 25: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/25.jpg)
25
Average Recall-Precision Graph Example. Average Interpolated recall-precision graph
(which is smoother than a single one) for 50 queries
![Page 26: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/26.jpg)
26
Focusing on Top Documents Users tend to look at only the top part of the ranked
result list to find relevant documents
Some search tasks have only one relevant document (P@1)
e.g., navigational (seen/assumed to exist) search, question answering
Recall not an appropriate measure in this case
Instead need to measure how well the search engine does at retrieving relevant documents at very high ranks
![Page 27: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/27.jpg)
27
Focusing on Top Documents Precision at Rank R
R typically is 5, 10, or 20
Easy to compute, average, understand
Not sensitive to rank positions less than R, higher or lower
Reciprocal Rank Reciprocal of the rank at which the 1st relevant document
is retrieved
Mean Reciprocal Rank (MRR) is the average of the reciprocal ranks over a set of queries
Very sensitive to rank position, e.g., dn, dr, dn, dn, dn (MRR = ½), whereas dn, dn, dn, dn, dr (MRR = 1/5)
![Page 28: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/28.jpg)
28
Discounted Cumulative Gain Popular measure for evaluating web search and related
tasks
Two assumptions:
Highly relevant documents are more useful than marginally relevant document
The lower the ranked position of a relevant document, the less useful it is for the user, since it is less likely to be examined
![Page 29: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/29.jpg)
29
Discounted Cumulative Gain Uses graded relevance (i.e., a value) as a measure of the
usefulness, or gain, from examining a document
Gain is accumulated starting at the top of the ranking and may be reduced, or discounted, at lower ranks
Typical discount is 1 / log(rank)
With base 2, the discount at rank 4 is , and at rank 8 it is
![Page 30: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/30.jpg)
30
Discounted Cumulative Gain DCG is the total gain accumulated at a particular rank p:
where reli is the graded relevance of the document at rank i, e.g., ranging from “Bad” to “Perfect” is 0 reli
5, or 0..1
Alternative formulation:
Used by some web search companies
Emphasis on retrieving highly relevant documents
log2 i is the discount/reduction factor applied to the gain
![Page 31: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/31.jpg)
31
DCG Example Assume that there are 10 ranked documents judged on a
0-3 (“not relevant” to “highly relevant”) relevance scale:
3, 2, 3, 0, 0, 1, 2, 2, 3, 0
which represent the gain at each rank
The discounted gain is (reli / log2 i, i 2)
3, 2/1, 3/1.59, 0, 0, 1/2.59, 2/2.81, 2/3, 3/3.17, 0
= 3, 2, 1.89, 0, 0, 0.39, 0.71, 0.67, 0.95, 0
DCG at each rank is formed by accumulating the numbers
3, 5, 6.89, 6.89, 6.89, 7.28, 7.99, 8.66, 9.61, 9.61(5) (10)
![Page 32: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/32.jpg)
32
Normalized DCG DCG numbers are averaged across a set of queries at
specific rank values, similar to precision@p
e.g., DCG at rank 5 is 6.89 and at rank 10 is 9.61
DCG values are often normalized by comparing the DCG at each rank with the DCG value for the perfect
ranking
NDCGp = DCGp / IDCGp
Makes averaging easier for queries with different numbers of relevant documents
![Page 33: Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines Measurement usually carried out.](https://reader037.fdocuments.in/reader037/viewer/2022110322/56649d2c5503460f94a01ed6/html5/thumbnails/33.jpg)
33
NDCG Example Perfect ranking for which each relevance level is on
the scale 0-3:
3, 3, 3, 2, 2, 2, 1, 0, 0, 0
Ideal DCG values:
3, 6, 7.89, 8.89, 9.75, 10.52, 10.88, 10.88, 10.88, 10
Given the DCG values are 3, 5, 6.89, 6.89, 6.89, 7.28, 7.99, 8.66, 9.61, 9.61, the NDCG
values:
1, 0.83, 0.87, 0.76, 0.71, 0.69, 0.73, 0.8, 0.88, 0.88
NDCG 1 at any rank position