Multi-modal fashion product retrieval
A Rubio Romano, LL Yu, E Simó Serra… - … 6th Workshop on …, 2017 - upcommons.upc.edu
Proceedings 6th Workshop on Vision and Language (VL), 2017•upcommons.upc.edu
Finding a product in the fashion world can be a daunting task. Everyday, e-commerce sites
are updating with thousands of images and their associated metadata (textual information),
deepening the problem. In this paper, we leverage both the images and textual metadata
and propose a joint multi-modal embedding that maps both the text and images into a
common latent space. Distances in the latent space correspond to similarity between
products, allowing us to effectively perform retrieval in this latent space. We compare against …
are updating with thousands of images and their associated metadata (textual information),
deepening the problem. In this paper, we leverage both the images and textual metadata
and propose a joint multi-modal embedding that maps both the text and images into a
common latent space. Distances in the latent space correspond to similarity between
products, allowing us to effectively perform retrieval in this latent space. We compare against …
Finding a product in the fashion world can be a daunting task. Everyday, e-commerce sites are updating with thousands of images and their associated metadata (textual information), deepening the problem. In this paper, we leverage both the images and textual metadata and propose a joint multi-modal embedding that maps both the text and images into a common latent space. Distances in the latent space correspond to similarity between products, allowing us to effectively perform retrieval in this latent space. We compare against existing approaches and show significant improvements in retrieval tasks on a largescale e-commerce dataset.
upcommons.upc.edu
Showing the best result for this search. See all results