Title | ||
---|---|---|
Squeezing Water from a Stone: A Bag of Tricks for Further Improving Cross-Encoder Effectiveness for Reranking |
Abstract | ||
---|---|---|
While much recent work has demonstrated that hard negative mining can be used to train better bi-encoder models, few have considered it in the context of cross-encoders, which are key ingredients in modern retrieval pipelines due to their high effectiveness. One noteworthy exception comes from Gao et al. [13], who propose to train cross-encoders by adapting the well-known NCE loss and augmenting it with a “localized” selection of hard negative examples from the first-stage retriever, which they call the Localized Contrastive Estimation (LCE) loss. In this work, we present a replication study of LCE on a different task and combine it with several other “tricks” (e.g., replacing
$$\text {BERT}_{\text {Base}}$$
with
$$\text {ELECTRA}_{\text {Base}}$$
and replacing BM25 with TCT-ColBERTv2) to substantially improve ranking effectiveness. We attempt to more systematically explore certain parts of the hyperparameter space, including the choice of losses and the group size in the LCE loss. While our findings, for the most part, align with those from the original paper, we observe that for MS MARCO passage, orienting the retriever used for hard negative mining with the first-stage retriever used for inference is not as critical for improving effectiveness across all settings. Our code and documentation can be found in:
https://github.com/castorini/replicate-lce
.
|
Year | DOI | Venue |
---|---|---|
2022 | 10.1007/978-3-030-99736-6_44 | Advances in Information Retrieval |
DocType | Volume | ISSN |
Conference | 13185 | 0302-9743 |
Citations | PageRank | References |
0 | 0.34 | 0 |
Authors | ||
6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Ronak Pradeep | 1 | 13 | 2.44 |
Yuqi Liu | 2 | 0 | 0.34 |
Xinyu Zhang | 3 | 1 | 1.36 |
Yilin Li | 4 | 0 | 0.68 |
Andrew Yates | 5 | 0 | 1.69 |
Jimmy Lin | 6 | 4800 | 376.93 |