Abstract | ||
---|---|---|
Axiomatic approaches aim to utilize reasonable retrieval constraints to guide the search for optimal retrieval models. Existing studies have shown the effectiveness of axiomatic approaches in improving the performance through either the derivation of new basic retrieval models or modifications of existing ones. Recently, neural network models have attracted more attention in the research community. Since these models are learned from training data, it would be interesting to study how to utilize the axiomatic approaches to guide the training process so that the learned models can satisfy retrieval constraints and achieve better retrieval performance.
In this paper, we propose to utilize axiomatic perturbations to construct training data sets for neural ranking models. The perturbed data sets are constructed in a way to amplify the desirable properties that any reasonable retrieval models should satisfy. As a result, the models learned from the perturbed data sets are expected to satisfy more retrieval constraints and lead to better retrieval performance. Experiment results show that the models learned from the perturbed data sets indeed perform better than those learned from the original data sets.
|
Year | DOI | Venue |
---|---|---|
2020 | 10.1145/3409256.3409828 | ICTIR '20: The 2020 ACM SIGIR International Conference on the Theory of Information Retrieval
Virtual Event
Norway
September, 2020 |
DocType | ISBN | Citations |
Conference | 978-1-4503-8067-6 | 1 |
PageRank | References | Authors |
0.35 | 0 | 2 |
Name | Order | Citations | PageRank |
---|---|---|---|
Zitong Cheng | 1 | 1 | 0.35 |
Hui Fang | 2 | 918 | 63.03 |