Title
Visualizing and Measuring the Geometry of BERT.
Abstract
Transformer architectures show significant promise for natural language processing. Given that a single pretrained model can be fine-tuned to perform well on many different tasks, these networks appear to extract generally useful linguistic features. How do such networks represent this information internally? This paper describes qualitative and quantitative investigations of one particularly effective model, BERT. At a high level, linguistic features seem to be represented in separate semantic and syntactic subspaces. We find evidence of a fine-grained geometric representation of word senses. We also present empirical descriptions of syntactic representations in both attention matrices and individual word embeddings, as well as a mathematical argument to explain the geometry of these representations.
Year
Venue
DocType
2019
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019)
Journal
Volume
ISSN
Citations 
32
1049-5258
0
PageRank 
References 
Authors
0.34
0
7
Name
Order
Citations
PageRank
Andy Coenen112.37
Emily Reif200.34
Ann Yuan341.86
Been Kim435321.44
Adam Pearce500.68
Fernanda B. Viégas600.34
Martin Wattenberg74695333.69