Title
A Comparison of Nuggets and Clusters for Evaluating Timeline Summaries.
Abstract
There is growing interest in systems that generate timeline summaries by filtering high-volume streams of documents to retain only those that are relevant to a particular event or topic. Continued advances in algorithms and techniques for this task depend on standardized and reproducible evaluation methodologies for comparing systems. However, timeline summary evaluation is still in its infancy, with competing methodologies currently being explored in international evaluation forums such as TREC. One area of active exploration is how to explicitly represent the units of information that should appear in a "good" summary. Currently, there are two main approaches, one based on identifying nuggets in an external "ground truth", and the other based on clustering system outputs. In this paper, by building test collections that have both nugget and cluster annotations, we are able to compare these two approaches. Specifically, we address questions related to evaluation effort, differences in the final evaluation products, and correlations between scores and rankings generated by both approaches. We summarize advantages and disadvantages of nuggets and clusters to offer recommendations for future system evaluations.
Year
DOI
Venue
2017
10.1145/3132847.3133000
CIKM
Field
DocType
ISBN
Data science,Data mining,Cluster (physics),Information retrieval,Computer science,Timeline,Ground truth,Cluster analysis
Conference
978-1-4503-4918-5
Citations 
PageRank 
References 
3
0.38
12
Authors
3
Name
Order
Citations
PageRank
Gaurav Baruah1245.68
Richard Mccreadie240332.43
Jimmy Lin34800376.93