Title
Fracture Mechanics Method for Word Embedding Generation of Neural Probabilistic Linguistic Model
Abstract
AbstractWord embedding, a lexical vector representation generated via the neural linguistic model NLM, is empirically demonstrated to be appropriate for improvement of the performance of traditional language model. However, the supreme dimensionality that is inherent in NLM contributes to the problems of hyperparameters and long-time training in modeling. Here, we propose a force-directed method to improve such problems for simplifying the generation of word embedding. In this framework, each word is assumed as a point in the real world; thus it can approximately simulate the physical movement following certain mechanics. To simulate the variation of meaning in phrases, we use the fracture mechanics to do the formation and breakdown of meaning combined by a 2-gram word group. With the experiments on the natural linguistic tasks of part-of-speech tagging, named entity recognition and semantic role labeling, the result demonstrated that the 2-dimensional word embedding can rival the word embeddings generated by classic NLMs, in terms of accuracy, recall, and text visualization.
Year
DOI
Venue
2016
10.1155/2016/3506261
Periodicals
Field
DocType
Volume
Computer science,Natural language processing,Artificial intelligence,Word embedding,Probabilistic logic,Language model,Visualization,Speech recognition,Vocabulary,Linguistics,Named-entity recognition,Machine learning,Semantics,Semantic role labeling
Journal
2016
Issue
ISSN
Citations 
1
1687-5265
0
PageRank 
References 
Authors
0.34
27
3
Name
Order
Citations
PageRank
Size Bi100.68
Xiao Liang200.34
Tinglei Huang33812.17