Evaluation of Image Retrieval Results

Download Report

Transcript Evaluation of Image Retrieval Results

Evaluation of Image Retrieval Results

 

Relevant

: images which meet user’s information need

Irrelevant

: images which don’t meet user’s information need Query: cat

Relevant Irrelevant

1

Accuracy

• Given a query, an engine classifies each image as “Relevant” or “Nonrelevant” Retrieved Not Retrieved Relevant tp fn Nonrelevant fp tn • • • The accuracy of an engine: the fraction of these classifications that are correct – (tp + tn) / ( tp + fp + fn + tn) Accuracy is a commonly used evaluation measure in machine learning classification work Why is this not a very useful evaluation measure in IR?

2

Why not just use accuracy?

• How to build a 99.9999% accurate search engine on a low budget….

Search for:

0 matching results found.

• People doing information retrieval want to find something and have a certain tolerance for junk.

3

Unranked retrieval evaluation: Precision and Recall

Precision

: fraction of retrieved images that are relevant 

Recall

: fraction of relevant image that are retrieved Retrieved Not Retrieved Relevant tp fn Precision: P = tp/(tp + fp) Recall: R = tp/(tp + fn) 4 Nonrelevant fp tn

Precision/Recall

   You can get high recall (but low precision ) by retrieving all images for all queries!

Recall is a non-decreasing function of the number of images retrieved In a good system, precision decreases as either the number of images retrieved or recall increases  This is not a theorem, but a result with strong empirical confirmation 5

A combined measure: F

• Combined measure that assesses precision/recall tradeoff is F measure (weighted harmonic mean):

F

  1

P

1  ( 1   ) 1

R

 (   2  1 )

PR

2

P

R

• • People usually use balanced F

1

– i.e., with  = 1 or  = ½ measure Harmonic mean is a conservative average – See CJ van Rijsbergen, Information Retrieval 6

Evaluating ranked results

• Evaluation of ranked results: – The system can return any number of results – By taking various numbers of the top returned images (levels of recall), the evaluator can produce a precision-recall curve 7

A precision-recall curve

1.0

0.8

0.6

0.4

0.2

0.0

0.0

0.2

0.4

Recall 0.6

0.8

1.0

8

Interpolated precision

• • Idea: If locally precision increases with increasing recall, then you should get to count that… So you take the max of precisions to right of value 10

A precision-recall curve

1.0

0.8

0.6

0.4

0.2

0.0

0.0

0.2

0.4

Recall 0.6

0.8

1.0

11

Summarizing a Ranking

• Graphs are good, but people want summary measures!

1. Precision and recall at fixed retrieval level • Precision-at-k: Precision of top k results • Recall-at-k: Recall of top k results • Perhaps appropriate for most of web search: all people want are good matches on the first one or two results pages • But: averages badly and has an arbitrary parameter of

k

12

Summarizing a Ranking

Summarizing a Ranking

2. Calculating precision at standard recall levels, from 0.0 to 1.0

– 11-point interpolated average precision • The standard measure in the early TREC competitions: you take the precision at 11 levels of recall varying from 0 to 1 by tenths of the images, using interpolation (the value for 0 is always interpolated!) , and average them • Evaluates performance at all recall levels 14

A precision-recall curve

1.0

0.8

0.6

0.4

0.2

0.0

0.0

0.2

0.4

Recall 0.6

0.8

1.0

15

Typical 11 point precisions

• SabIR/Cornell 8A1 11pt precision from TREC 8 (1999) 1 0.8

0.6

0.4

0.2

0 0 0.2

0.4

Recall

0.6

0.8

1 16

Summarizing a Ranking

3. Average precision (AP) – Averaging the precision values from the rank positi ons where a relevant image was retrieved – Avoids interpolation, use of fixed recall levels – MAP for query collection is arithmetic average • Macro-averaging: each query counts equally 17

Average Precision

18

Summarizing a Ranking

3. Mean average precision (MAP) – summarize rankings from multiple queries by aver aging average precision – most commonly used measure in research papers – assumes user is interested in finding many relevan t images for each query 19

Summarizing a Ranking

4. R-precision – If we have a known (though perhaps incomplete) set of relevant images of size Rel, then calculate precision of the top Rel images returned – Perfect system could score 1.0.

21

Summarizing a Ranking for Multiple Relevance levels

Query Van gogh paintings Excellent Relevant Irrelevant Key ideas

- painting & good quality

Bridge Hugh Grant

- Can see full bridge, visually pleasing - picture of one person verse group or no person - picture of head, clear image, good quality 22

Summarizing a Ranking for Multiple Relevance levels 5. NDCG: Normalized Discounted Cumulative Gain – Popular measure for evaluating web search and re lated tasks – Two assumptions: • Highly relevant images are more useful than marginally relevant image • the lower the ranked position of a relevant image, the l ess useful it is for the user, since it is less likely to be exa mined 23

Summarizing a Ranking for Multiple Relevance levels 5. DCG: Discounted Cumulative Gain – the total gain accumulated at a particular rank p: – Alternative formulation • emphasis on retrieving highly relevant images 24

Summarizing a Ranking for Multiple Relevance levels 5. DCG: Discounted Cumulative Gain – 10 ranked images judged on 0‐3 relevance scale: 3, 2, 3, 0, 0, 1, 2, 2, 3, 0 – discounted gain: 3, 2/1, 3/1.59, 0, 0, 1/2.59, 2/2.81, 2/3, 3/3.17, 0 = 3, 2, 1.89, 0, 0, 0.39, 0.71, 0.67, 0.95, 0 – DCG: 3, 5, 6.89, 6.89, 6.89, 7.28, 7.99, 8.66, 9.61, 9.61

25

Summarizing a Ranking for Multiple Relevance levels 5. NDCG – DCG values are often normalized by comparing the DCG at each rank with the DCG value for the perfect ranking • makes averaging easier for queries with different numbers of r elevant images – Perfect ranking: 3, 3, 3, 2, 2, 2, 1, 0, 0, 0 – ideal DCG values: 3, 6, 7.89, 8.89, 9.75, 10.52, 10.88, 10.88, 10.88, 10 – NDCG values (divide actual by ideal): 1, 0.83, 0.87, 0.76, 0.71, 0.69, 0.73, 0.8, 0.88, 0.88

NDCG ≤1 at any rank position 26

Variance

• • • For a test collection, it is usual that a system does crummily on some information needs (e.g., MAP = 0.1) and excellently on others (e.g., MAP = 0.7) Indeed, it is usually the case that the variance in performance of the same system across queries is much greater than the variance of different systems on the same query.

That is, there are easy information needs and hard ones!

27

Significance Tests

• • Given the results from a number of queries, how can we conclude that ranking algorithm A is better than a lgorithm B?

A significance test enables us to reject the null hypot hesis (no difference) in favor of the alternative hypot hesis (B is better than A) – the power of a test is the probability that the test will rejec t the null hypothesis correctly – increasing the number of queries in the experime nt also increases power of test 28