Abstract | ||
---|---|---|
Word embeddings have been found useful for many NLP tasks, including part-of-speech tagging, named entity recognition, and parsing. Adding multilingual context when learning embeddings can improve their quality, for example via canonical correlation analysis (CCA) on embeddingsfromtwo languages. In this paper, we extend this idea to learn deep non-linear transformations of word embeddings of the two languages, using the recently proposed deep canonical correlation analysis. The resulting embeddings, when evaluated on multiple word and bigram similarity tasks, consistently improve over monolingual embeddings and over embeddings transformed with linear CCA. |
Year | Venue | Field |
---|---|---|
2015 | HLT-NAACL | Computer science,Canonical correlation,Speech recognition,Correlation,Bigram,Artificial intelligence,Natural language processing,Parsing,Named-entity recognition |
DocType | Citations | PageRank |
Conference | 39 | 1.07 |
References | Authors | |
34 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Ang Lu | 1 | 40 | 1.42 |
Weiran Wang | 2 | 114 | 9.99 |
Mohit Bansal | 3 | 871 | 63.19 |
Kevin Gimpel | 4 | 1545 | 79.71 |
Karen Livescu | 5 | 1254 | 71.43 |