Title
Learning Phrase Representations Based on Word and Character Embeddings.
Abstract
Most phrase embedding methods consider a phrase as a basic term and learn embeddings according to phrases' external contexts, ignoring the internal structures of words and characters. There are some languages such as Chinese, a phrase is usually composed of several words or characters and contains rich internal information. The semantic meaning of a phrase is also related to the meanings of its composing words or characters. Therefore, we take Chinese for example, and propose a joint words and characters embedding model for learning phrase representation. In order to disambiguate the word and character and address the issue of non-compositional phrases, we present multiple-prototype word and character embeddings and an effective phrase selection method. We evaluate the effectiveness of the proposed model on phrase similarities computation and analogical reasoning. The empirical result shows that our model outperforms other baseline methods which ignore internal word and character information.
Year
DOI
Venue
2016
10.1007/978-3-319-46681-1_65
Lecture Notes in Computer Science
Keywords
Field
DocType
Embedding,Phrase representation,Semantic composition,Analogical reasoning
Noun phrase,Analogical reasoning,Embedding,Computer science,Phrase,Natural language processing,Artificial intelligence,Computation
Conference
Volume
ISSN
Citations 
9950
0302-9743
1
PageRank 
References 
Authors
0.35
11
5
Name
Order
Citations
PageRank
Jiangping Huang110.68
Donghong Ji221.03
Shuxin Yao310.68
Wen-Zhi Huang4132.92
Bo Chen521.38