Title
Swivel: Improving Embeddings by Noticing What's Missing.
Abstract
We present Submatrix-wise Vector Embedding Learner (Swivel), a method for generating low-dimensional feature embeddings from a feature co-occurrence matrix. Swivel performs approximate factorization of the point-wise mutual information matrix via stochastic gradient descent. It uses a piecewise loss with special handling for unobserved co-occurrences, and thus makes use of all the information in the matrix. While this requires computation proportional to the size of the entire matrix, we make use of vectorized multiplication to process thousands of rows and columns at once to compute millions of predicted values. Furthermore, we partition the matrix into shards in order to parallelize the computation across many nodes. This approach results in more accurate embeddings than can be achieved with methods that consider only observed co-occurrences, and can scale to much larger corpora than can be handled with sampling methods.
Year
Venue
Field
2016
arXiv: Computation and Language
Row and column spaces,Stochastic gradient descent,Embedding,Matrix (mathematics),Computer science,Theoretical computer science,Mutual information,Artificial intelligence,Factorization,Machine learning,Piecewise,Computation
DocType
Volume
Citations 
Journal
abs/1602.02215
11
PageRank 
References 
Authors
0.55
11
4
Name
Order
Citations
PageRank
Noam Shazeer1108943.70
Ryan Doherty2131.25
Colin Evans3131.25
Chris Waterson4120.90