Abstract | ||
---|---|---|
Most phrase embedding methods consider a phrase as a basic term and learn embeddings according to phrases' external contexts, ignoring the internal structures of words and characters. There are some languages such as Chinese, a phrase is usually composed of several words or characters and contains rich internal information. The semantic meaning of a phrase is also related to the meanings of its composing words or characters. Therefore, we take Chinese for example, and propose a joint words and characters embedding model for learning phrase representation. In order to disambiguate the word and character and address the issue of non-compositional phrases, we present multiple-prototype word and character embeddings and an effective phrase selection method. We evaluate the effectiveness of the proposed model on phrase similarities computation and analogical reasoning. The empirical result shows that our model outperforms other baseline methods which ignore internal word and character information. |
Year | DOI | Venue |
---|---|---|
2016 | 10.1007/978-3-319-46681-1_65 | Lecture Notes in Computer Science |
Keywords | Field | DocType |
Embedding,Phrase representation,Semantic composition,Analogical reasoning | Noun phrase,Analogical reasoning,Embedding,Computer science,Phrase,Natural language processing,Artificial intelligence,Computation | Conference |
Volume | ISSN | Citations |
9950 | 0302-9743 | 1 |
PageRank | References | Authors |
0.35 | 11 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Jiangping Huang | 1 | 1 | 0.68 |
Donghong Ji | 2 | 2 | 1.03 |
Shuxin Yao | 3 | 1 | 0.68 |
Wen-Zhi Huang | 4 | 13 | 2.92 |
Bo Chen | 5 | 2 | 1.38 |