Title
SoftRank: optimizing non-smooth rank metrics
Abstract
We address the problem of learning large complex ranking functions. Most IR applications use evaluation metrics that depend only upon the ranks of documents. However, most ranking functions generate document scores, which are sorted to produce a ranking. Hence IR metrics are innately non-smooth with respect to the scores, due to the sort. Unfortunately, many machine learning algorithms require the gradient of a training objective in order to perform the optimization of the model parameters,and because IR metrics are non-smooth,we need to find a smooth proxy objective that can be used for training. We present a new family of training objectives that are derived from the rank distributions of documents, induced by smoothed scores. We call this approach SoftRank. We focus on a smoothed approximation to Normalized Discounted Cumulative Gain (NDCG), called SoftNDCG and we compare it with three other training objectives in the recent literature. We present two main results. First, SoftRank yields a very good way of optimizing NDCG. Second, we show that it is possible to achieve state of the art test set NDCG results by optimizing a soft NDCG objective on the training set with a different discount function
Year
DOI
Venue
2008
10.1145/1341531.1341544
WSDM
Keywords
Field
DocType
cumulant,machine learning,metrics,ranking,optimization,gradient descent
Training set,Data mining,Learning to rank,Gradient descent,Discount function,Ranking,Ranking SVM,Computer science,sort,Artificial intelligence,Machine learning,Test set
Conference
Citations 
PageRank 
References 
136
5.11
12
Authors
4
Search Limit
100136
Name
Order
Citations
PageRank
Michael J. Taylor174941.75
John Guiver248221.48
STEPHEN ROBERTSON36204669.07
Tom Minka472039.57