Title
Pretraining Multi-modal Representations for Chinese NER Task with Cross-Modality Attention
Abstract
ABSTRACTNamed Entity Recognition (NER) aims to identify the pre-defined entities from the unstructured text. Compared with English NER, Chinese NER faces more challenges: the ambiguity problem in entity boundary recognition due to unavailable explicit delimiters between Chinese characters, and the out-of-vocabulary (OOV) problem caused by rare Chinese characters. However, two important features specific to the Chinese language are ignored by previous studies: glyphs and phonetics, which contain rich semantic information of Chinese. To overcome these issues by exploiting the linguistic potential of Chinese as a logographic language, we present MPM-CNER (short for Multi-modal Pretraining Model for Chinese NER), a model for learning multi-modal representations of Chinese semantics, glyphs, and phonetics, via four pretraining tasks: Radical Consistency Identification (RCI), Glyph Image Classification (GIC), Phonetic Consistency Identification (PCI), and Phonetic Classification Modeling (PCM). Meanwhile, a novel cross-modality attention mechanism is proposed to fuse these multimodal features for further improvement. The experimental results show that our method outperforms the state-of-the-art baseline methods on four benchmark datasets, and the ablation study also verifies the effectiveness of the pre-trained multi-modal representations.
Year
DOI
Venue
2022
10.1145/3488560.3498450
WSDM
Keywords
DocType
Citations 
Chinese named entity recognition, multi-modal representations, pre-training model, cross-modality attention
Conference
0
PageRank 
References 
Authors
0.34
0
7
Name
Order
Citations
PageRank
Chengcheng Mai100.68
Mengchuan Qiu201.35
Kaiwen Luo300.68
Ziyan Peng400.68
J. Liu56415.00
Chunfeng Yuan656.90
Yihua Huang786.61