Unit 3 - Evaluation And Visualization Of IR Flashcards
What is performance evaluation?
The measure of how much the retrieved documents are relevant to the users query.
Categories of the results obtained from a information retrieval system:
Relevant and retrieved
Relevant and not retrieved
Non-relevant and retrieved
Non-relevant and not retrieved
What are relevant items?
Document which are actually useful to the user.
What is precision?
Ratio of number of relevant and retrieved documents with the total number of retrieved documents
P = (A / (A + C))x100%
What is recall?
Ratio of number of relevant and retrieve documents to the total number of relevant documents in the database.
R = (A /(A + B))x100%
Problem with precision and recall
They are inversely proportional to each other
Need of detailed knowledge about The relevant items in the database
Mean reciprocal rank (MRR)
It is the rank aware evaluation metric i.e. it considers the relevance as well as the rank of the documents
Comprises of binary relevance based metric
Algorithm of MRR:
What is F-measure?
Measure that combines the recall and precision.
Also known as harmonic mean.
Normalise discounted cumulative gain (NDCG)
It is a measure of ranking quality.
Assumption that needs to be kept in mind for NDCG
Highly relevant documents are more useful than moderately relevant documents which are in turn more useful than irrelevant documents 
To find NDCG, calculate
Cumulative gain
Discounted cumulative gain
Ideal discounted communicative gain
Normalised discounted cumulative gain