Title
Wan2vec: Embeddings learned on word association norms.
Abstract
Word embeddings are powerful for many tasks in natural language processing. In this work, we learn word embeddings using weighted graphs from word association norms (WAN) with the node2vec algorithm. Although building WAN is a difficult and time-consuming task, training the vectors from these resources is a fast and efficient process. This allows us to obtain good quality word embeddings from small corpora. We evaluate our word vectors in two ways: intrinsic and extrinsic. The intrinsic evaluation was performed with several word similarity benchmarks, WordSim-353 , MC30, MTurk-287, MEN-TR-3k, SimLex-999, MTurk-771 and RG-65 , and different similarity measures achieving better results than those obtained with word2vec, GloVe, and fastText, trained on a huge corpus. The extrinsic evaluation was done by measuring the quality of sentence embeddings using transfer tasks: sentiment analysis, paraphrase detection, natural language inference, and semantic textual similarity. The word vectors learned from the WAN are available on our Github page.
Year
DOI
Venue
2019
10.3233/SW-190349
SEMANTIC WEB
Keywords
DocType
Volume
Word association norms,word embeddings,word similarity,word2vec,GloVe,fastText
Journal
10
Issue
ISSN
Citations 
6
1570-0844
0
PageRank 
References 
Authors
0.34
0
4
Name
Order
Citations
PageRank
Gemma Bel-Enguix100.34
Helena Gómez-Adorno24016.01
Jorge Reyes-Magaña300.68
Gerardo Sierra47822.35