Abstract
The key challenge for learning a fine-grained sketch-based image retrieval (FG-SBIR) model is to bridge the domain gap between photo and sketch. Existing models learn a deep joint embedding space with discriminative losses where a photo and a sketch can be compared. In this paper, we propose a novel discriminative-generative hybrid model by introducing a generative task of cross-domain image synthesis. This task enforces the learned embedding space to preserve all the domain invariant information that is useful for cross-domain reconstruction, thus explicitly reducing the domain gap as opposed to existing models. Extensive experiments on the largest FG-SBIR dataset Sketchy [19] show that the proposed model significantly outperforms state-of-the-art discriminative FG-SBIR models.
Original language | English |
---|---|
Title of host publication | The British Machine Vision Conference (BMVC 2017) |
Number of pages | 12 |
ISBN (Electronic) | 1-901725-60-X |
DOIs | |
Publication status | E-pub ahead of print - 7 Sept 2017 |
Event | The 28th British Machine Vision Conference - Imperial College London, London, United Kingdom Duration: 4 Sept 2017 → 7 Sept 2017 https://bmvc2017.london/ |
Conference
Conference | The 28th British Machine Vision Conference |
---|---|
Abbreviated title | BMVC 2017 |
Country/Territory | United Kingdom |
City | London |
Period | 4/09/17 → 7/09/17 |
Internet address |