Title | ||
---|---|---|
Phrase embedding learning based on external and internal context with compositionality constraint. |
Abstract | ||
---|---|---|
Different methods are proposed to learn phrase embedding, which can be mainly divided into two strands. The first strand is based on the distributional hypothesis to treat a phrase as one non-divisible unit and to learn phrase embedding based on its external context similar to learn word embedding. However, distributional methods cannot make use of the information embedded in component words and they also face data spareness problem. The second strand is based on the principle of compositionality to infer phrase embedding based on the embedding of its component words. Compositional methods would give erroneous result if a phrase is non-compositional. In this paper, we propose a hybrid method by a linear combination of the distributional component and the compositional component with an individualized phrase compositionality constraint. The phrase compositionality is automatically computed based on the distributional embedding of the phrase and its component words. Evaluation on five phrase level semantic tasks and experiments show that our proposed method has overall best performance. Most importantly, our method is more robust as it is less sensitive to datasets. |
Year | DOI | Venue |
---|---|---|
2018 | 10.1016/j.knosys.2018.04.009 | Knowledge-Based Systems |
Keywords | Field | DocType |
Phrase embedding,Compositionality,Distributional hypothesis,Composition model | Principle of compositionality,Linear combination,Embedding,Computer science,Phrase,Artificial intelligence,Natural language processing,Word embedding,Machine learning | Journal |
Volume | Issue | ISSN |
152 | C | 0950-7051 |
Citations | PageRank | References |
0 | 0.34 | 38 |
Authors | ||
4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Li Minglei | 1 | 44 | 12.21 |
Qin Lu | 2 | 15 | 9.47 |
Dan Xiong | 3 | 3 | 3.73 |
Yunfei Long | 4 | 2 | 1.72 |