Abstract | ||
---|---|---|
Metaphorical expressions are pervasive in natural language and pose a substantial challenge for computational semantics. The inherent compositionality of metaphor makes it an important test case for compositional distributional semantic models (CDSMs). This paper is the first to investigate whether metaphorical composition warrants a distinct treatment in the CDSM framework. We propose a method to learn metaphors as linear transformations in a vector space and find that, across a variety of semantic domains, explicitly modeling metaphor improves the resulting semantic representations. We then use these representations in a metaphor identification task, achieving a high performance of 0.82 in terms of F-score. |
Year | Venue | DocType |
---|---|---|
2016 | PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1 | Conference |
Volume | Citations | PageRank |
P16-1 | 5 | 0.44 |
References | Authors | |
22 | 4 |
Name | Order | Citations | PageRank |
---|---|---|---|
E. Gutiérrez | 1 | 20 | 3.88 |
Ekaterina Shutova | 2 | 228 | 21.51 |
Tyler Marghetis | 3 | 6 | 7.30 |
benjamin bergen | 4 | 12 | 6.38 |