Title
Multi-modal joint embedding for fashion product retrieval
Abstract
Finding a product in the fashion world can be a daunting task. Everyday, e-commerce sites are updating with thousands of images and their associated metadata (textual information), deepening the problem, akin to finding a needle in a haystack. In this paper, we leverage both the images and textual metadata and propose a joint multi-modal embedding that maps both the text and images into a common latent space. Distances in the latent space correspond to similarity between products, allowing us to effectively perform retrieval in this latent space, which is both efficient and accurate. We train this embedding using large-scale real world e-commerce data by both minimizing the similarity between related products and using auxiliary classification networks to that encourage the embedding to have semantic meaning. We compare against existing approaches and show significant improvements in retrieval tasks on a large-scale e-commerce dataset. We also provide an analysis of the different metadata.
Year
DOI
Venue
2017
10.1109/ICIP.2017.8296311
2017 IEEE International Conference on Image Processing (ICIP)
Keywords
Field
DocType
Multi-modal embedding,neural networks,retrieval
Metadata,Active vision,Embedding,Haystack,Task analysis,Pattern recognition,Information retrieval,Computer science,Artificial intelligence,Deep learning,Semantics,Modal
Conference
ISSN
ISBN
Citations 
1522-4880
978-1-5090-2176-5
0
PageRank 
References 
Authors
0.34
19
4
Name
Order
Citations
PageRank
Antonio Rubio161.92
LongLong Yu231.41
Edgar Simo-Serra364627.31
Francesc Moreno-Noguer4164793.46