A nested heuristic for parameter tuning in Support Vector Machines

E. Carrizosa, Belen Martin-Barragan, Dolores Romero Morales

Research output: Contribution to journalArticlepeer-review

Abstract / Description of output

The default approach for tuning the parameters of a Support Vector Machine (SVM) is a grid search in the parameter space. Different metaheuristics have been recently proposed as a more efficient alternative, but they have only shown to be useful in models with a low number of parameters. Complex models, involving many parameters, can be seen as extensions of simpler and easy-to-tune models, yielding a nested sequence of models of increasing complexity. In this paper we propose an algorithm which successfully exploits this nested property, with two main advantages versus the state of the art. First, our framework is general enough to allow one to address, with the very same method, several popular SVM parameter models encountered in the literature. Second, as algorithmic requirements we only need either an SVM library or any routine for the minimization of convex quadratic functions under linear constraints. In the computational study, we address Multiple Kernel Learning tuning problems for which grid search clearly would be infeasible, while our classification accuracy is comparable to that of ad hoc model-dependent benchmark tuning methods.
Original languageEnglish
Pages (from-to)328-334
JournalComputers and Operations Research
Volume43
Early online date15 Oct 2013
DOIs
Publication statusPublished - Mar 2014

Keywords / Materials (for Non-textual outputs)

  • Supervised classification
  • Support Vector Machines
  • Multiple kernel learning
  • Parameter tuning
  • Nested heuristic; Variable neighborhood search

Fingerprint

Dive into the research topics of 'A nested heuristic for parameter tuning in Support Vector Machines'. Together they form a unique fingerprint.

Cite this