Title
A Reproducibility Study of Information Retrieval Models.
Abstract
Developing effective information retrieval models has been a long standing challenge in Information Retrieval (IR), and significant progresses have been made over the years. With the increasing number of developed retrieval functions and the release of new data collections, it becomes more difficult, if not impossible, to compare a new retrieval function with all existing retrieval functions over all available data collections. To tackle thisproblem, this paper describes our efforts on constructing a platform that aims to improve the reproducibility of IR researchand facilitate the evaluation and comparison of retrieval functions. With the developed platform, more than 20 state of the art retrieval functions have been implemented and systematically evaluated over 16 standard TREC collections (including the newly released ClueWeb datasets). Our reproducibility study leads to several interesting observations. First, the performance difference between the reproduced results and those reported in the original papers is small for most retrieval functions. Second, the optimal performance of a few representative retrieval functions is still comparable over the new TREC ClueWeb collections. Finally, the developed platform (i.e., RISE ) is made publicly available so that any IR researchers would be able to utilize it to evaluate other retrieval functions.
Year
Venue
Field
2016
ICTIR
Data mining,Evaluation system,Human–computer information retrieval,Information retrieval,Computer science,Data retrieval,Relevance (information retrieval)
DocType
Citations 
PageRank 
Conference
4
0.44
References 
Authors
23
2
Name
Order
Citations
PageRank
Peilin Yang110012.00
Hui Fang291863.03