Abstract
In this paper we introduce a sparse kernel learning framework for the Continuous Relevance Model (CRM). State-of-the-art image annotation models linearly combine evidence from several different feature types to improve image annotation accuracy. While previous authors have focused on learning the linear combination weights for these features, there has been no work examining the optimal combination of kernels. We address this gap by formulating a sparse kernel learning framework for the CRM, dubbed the SKL-CRM, that greedily selects an optimal combination of kernels. Our kernel learning framework rapidly converges to an annotation accuracy that substantially outperforms a host of state-of-the-art annotation models. We make two surprising conclusions: firstly, if the kernels are chosen correctly, only a very small number of features are required so to achieve superior performance over models that utilise a full suite of feature types; and secondly, the standard default selection of kernels commonly used in the literature is sub-optimal, and it is much better to adapt the kernel choice based on the feature type and image dataset.
Original language | English |
---|---|
Title of host publication | Proceedings of International Conference on Multimedia Retrieval |
Place of Publication | New York, NY, USA |
Publisher | ACM |
Number of pages | 8 |
ISBN (Print) | 978-1-4503-2782-4 |
DOIs | |
Publication status | Published - 2014 |
Keywords / Materials (for Non-textual outputs)
- Image Annotation, Statistical Models, Visual Features