Abstract
Face hallucination has been well studied in the last decade because of its useful applications in law enforcement and entertainment. Promising results on the problem of sketch-photo face hallucination have been achieved with classic, and increasingly deep learning-based methods. However, synthesized photos still lack the crisp fidelity of real photos. More importantly, good results have primarily been demonstrated on very constrained datasets where the style variability is very low, and crucially the sketches are perfectly align-able traces of the ground-truth photos. However, realistic applications in entertainment or law enforcement require working with more unconstrained sketches drawn from memory or description, which are not rigidly align-able. In this paper, we develop a new deep learning approach to address these settings. Our image-image regression network is trained with a combination of content and adversarial losses to generate crisp photorealistic images, and it contains an integrated spatial transformer network to deal with non-rigid alignment between the domains. We evaluate face synthesis on classic constrained, as well as unviewed, benchmarks namely CUHK, MGDB, and FSMD. The results qualitatively and quantitatively outperform existing approaches.
Original language | English |
---|---|
Title of host publication | The British Machine Vision Conference (BMVC 2017) |
Number of pages | 11 |
ISBN (Electronic) | 1-901725-60-X |
DOIs | |
Publication status | E-pub ahead of print - 7 Sep 2017 |
Event | The 28th British Machine Vision Conference - Imperial College London, London, United Kingdom Duration: 4 Sep 2017 → 7 Sep 2017 https://bmvc2017.london/ |
Conference
Conference | The 28th British Machine Vision Conference |
---|---|
Abbreviated title | BMVC 2017 |
Country/Territory | United Kingdom |
City | London |
Period | 4/09/17 → 7/09/17 |
Internet address |