Title
Evaluating non-deterministic retrieval systems
Abstract
The use of sampling, randomized algorithms, or training based on the unpredictable inputs of users in Information Retrieval often leads to non-deterministic outputs. Evaluating the effectiveness of systems incorporating these methods can be challenging since each run may produce different effectiveness scores. Current IR evaluation techniques do not address this problem. Using the context of distributed information retrieval as a case study for our investigation, we propose a solution based on multivariate linear modeling. We show that the approach provides a consistent and reliable method to compare the effectiveness of non-deterministic IR algorithms, and explain how statistics can safely be used to show that two IR algorithms have equivalent effectiveness.
Year
DOI
Venue
2014
10.1145/2600428.2609472
SIGIR
Keywords
Field
DocType
experimentation,statistical analysis,information retrieval,effectiveness evaluation,measurement,experimental design,performance evaluation
Randomized algorithm,Data mining,IR evaluation,Linear model,Computer science,Multivariate statistics,Sampling (statistics),Artificial intelligence,Machine learning,Statistical analysis
Conference
Citations 
PageRank 
References 
3
0.49
5
Authors
5
Name
Order
Citations
PageRank
Gaya K. Jayasinghe1202.18
William Webber258531.30
Mark Sanderson33751341.56
Lasitha S. Dharmasena461.25
Shane Culpepper551947.52