Title
Chinese keyword extraction model with distributed computing
Abstract
The content and structure of the news text are relatively complex and cannot effectively capture the core content. Existing supervision models cannot achieve good results in areas with less annotated data. To this end, we propose a new Chinese keyword extraction model with a distributed computing method. Precisely, we first fused the Bidirectional Encoder Representation from Transformers (BERT) and Conditional Random Fields (CRF) so that each word learns its relationship with the context while reducing errors; secondly, the adversarial training encourages the model to retain a small amount of annotations Sample knowledge to help extract keywords from unannotated samples; and because the model contains a large number of time-consuming components, it creatively uses distributed computing to save overall computing time. The results show that our model can steadily improve the performance of keyword phrase extraction in areas with insufficient labeled samples.
Year
DOI
Venue
2022
10.1016/j.compeleceng.2021.107639
COMPUTERS & ELECTRICAL ENGINEERING
Keywords
DocType
Volume
Keyword Extraction, BERT, Adversarial training, CRF, Distributed computing
Journal
97
ISSN
Citations 
PageRank 
0045-7906
0
0.34
References 
Authors
0
6
Name
Order
Citations
PageRank
Tiantian Ding100.34
Wenzhong Yang200.68
Fuyuan Wei300.34
Chao Ding400.34
Peng Kang500.34
Wenxiu Bu600.34