Abstract | ||
---|---|---|
ABSTRACTExpert Finding is an important task in Community Question Answering (CQA) platforms, which could help route questions to potential expertise users to answer. The key is to model the question content and experts based on their historical answered questions accurately. Recently Pretrained Language Models (PLMs, e.g., Bert) have shown superior text modeling ability and have been used in expert finding preliminary. However, most PLMs-based models focus on the corpus or document granularity during pretraining, which is inconsistent with the downstream expert modeling and finding task. In this paper, we propose an expert-level pretraining language model named ExpertBert, aiming to model questions, experts as well as question-expert matching effectively in a pretraining manner. In our approach, we aggregate the historical answered questions of an expert as the expert-specific input.Besides, we integrate the target question into the input and design a label-augmented Masked Language Model (MLM) task to further capture the matching pattern between question and experts, which makes the pretraining objectives that more closely resemble the downstream expert finding task. Experimental results and detailed analysis on real-world CQA datasets demonstrate the effectiveness of our ExpertBert. |
Year | DOI | Venue |
---|---|---|
2022 | 10.1145/3511808.3557597 | Conference on Information and Knowledge Management |
DocType | Citations | PageRank |
Conference | 0 | 0.34 |
References | Authors | |
0 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Hongtao Liu | 1 | 0 | 0.34 |
Zhepeng Lv | 2 | 0 | 0.34 |
Qing Yang | 3 | 0 | 1.35 |
Dongliang Xu | 4 | 0 | 0.68 |
Qiyao Peng | 5 | 0 | 0.34 |