Abstract / Description of output
Contemporary deep learning techniques have made image recognition a reasonably reliable technology. However training effective photo classifiers typically takes numerous examples which limits image recognition’s scalability and applicability to scenarios where images may not be available. This has motivated investigation into zero-shot learning, which addresses the issue via knowledge transfer from other modalities such as text. In this paper we investigate an alternative approach of synthesizing image classifiers: almost directly from a user’s imagination, via freehand sketch. This approach doesn’t require the category to be nameable or describable via attributes as per zero-shot learning. We achieve this via training a model regression network to map from free-hand sketch space to the space of photo classifiers. It turns out that this mapping can be learned in a category-agnostic way, allowing photo classifiers for new categories to be synthesized by user with no need for annotated training photos. We also demonstrate that this modality of classifier generation can also be used to enhance the granularity of an existing photo classifier, or as a complement to name-based zero-shot learning.
Original language | English |
---|---|
Title of host publication | Computer Vision and Pattern Recognition 2018 |
Publisher | Institute of Electrical and Electronics Engineers (IEEE) |
Number of pages | 9 |
Publication status | Published - 17 Dec 2018 |
Event | Computer Vision and Pattern Recognition 2018 - Salt Lake City, United States Duration: 18 Jun 2018 → 22 Jun 2018 http://cvpr2018.thecvf.com/ http://cvpr2018.thecvf.com/ http://cvpr2018.thecvf.com/ |
Publication series
Name | |
---|---|
Publisher | IEEE |
ISSN (Print) | 1063-6919 |
ISSN (Electronic) | 2575-7075 |
Conference
Conference | Computer Vision and Pattern Recognition 2018 |
---|---|
Abbreviated title | CVPR 2018 |
Country/Territory | United States |
City | Salt Lake City |
Period | 18/06/18 → 22/06/18 |
Internet address |