Localized algorithms for multiple kernel learning

Mehmet Gönen, Ethem Alpaydin

Research output: Contribution to journalArticlepeer-review

72 Scopus citations

Abstract

Instead of selecting a single kernel, multiple kernel learning (MKL) uses a weighted sum of kernels where the weight of each kernel is optimized during training. Such methods assign the same weight to a kernel over the whole input space, and we discuss localized multiple kernel learning (LMKL) that is composed of a kernel-based learning algorithm and a parametric gating model to assign local weights to kernel functions. These two components are trained in a coupled manner using a two-step alternating optimization algorithm. Empirical results on benchmark classification and regression data sets validate the applicability of our approach. We see that LMKL achieves higher accuracy compared with canonical MKL on classification problems with different feature representations. LMKL can also identify the relevant parts of images using the gating model as a saliency detector in image recognition problems. In regression tasks, LMKL improves the performance significantly or reduces the model complexity by storing significantly fewer support vectors.

Original languageEnglish (US)
Pages (from-to)795-807
Number of pages13
JournalPattern Recognition
Volume46
Issue number3
DOIs
StatePublished - Mar 2013
Externally publishedYes

Keywords

  • Classification
  • Multiple kernel learning
  • Regression
  • Selective attention
  • Support vector machines
  • Support vector regression

ASJC Scopus subject areas

  • Software
  • Signal Processing
  • Computer Vision and Pattern Recognition
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'Localized algorithms for multiple kernel learning'. Together they form a unique fingerprint.

Cite this