A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2018; you can also visit the original URL.
The file type is application/pdf
.
Multi-modal joint embedding for fashion product retrieval
2017
2017 IEEE International Conference on Image Processing (ICIP)
Finding a product in the fashion world can be a daunting task. Everyday, e-commerce sites are updating with thousands of images and their associated metadata (textual information), deepening the problem, akin to finding a needle in a haystack. In this paper, we leverage both the images and textual metadata and propose a joint multi-modal embedding that maps both the text and images into a common latent space. Distances in the latent space correspond to similarity between products, allowing us
doi:10.1109/icip.2017.8296311
dblp:conf/icip/RubioYSM17
fatcat:wi4onryz6rcfxnc2e5jcrslzym