Title
Deep Embedding for Spatial Role Labeling.
Abstract
This paper introduces the visually informed embedding of word (VIEW), a continuous vector representation for a word extracted from a deep neural model trained using the Microsoft COCO data set to forecast the spatial arrangements between visual objects, given a textual description. The model is composed of a deep multilayer perceptron (MLP) stacked on the top of a Long Short Term Memory (LSTM) network, the latter being preceded by an embedding layer. The VIEW is applied to transferring multimodal background knowledge to Spatial Role Labeling (SpRL) algorithms, which recognize spatial relations between objects mentioned in the text. This work also contributes with a new method to select complementary features and a fine-tuning method for MLP that improves the $F1$ measure in classifying the words into spatial roles. The VIEW is evaluated with the Task 3 of SemEval-2013 benchmark data set, SpaceEval.
Year
Venue
Field
2016
arXiv: Computation and Language
Spatial relation,Visual Objects,Embedding,Pattern recognition,Computer science,Long short term memory,Multilayer perceptron,Natural language processing,Artificial intelligence,Machine learning
DocType
Volume
Citations 
Journal
abs/1603.08474
4
PageRank 
References 
Authors
0.44
9
4
Name
Order
Citations
PageRank
Oswaldo Ludwig1894.97
Xiao Liu299284.21
Parisa Kordjamshidi314318.52
Marie-Francine Moens41750139.27