Updating kernel methods in spectral decomposition by affinity perturbations

Yaniv Shmueli, Guy Wolf, Amir Averbuch

Research output: Contribution to journalArticlepeer-review


Many machine learning based algorithms contain a training step that is done once. The training step is usually computational expensive since it involves processing of huge matrices. If the training profile is extracted from an evolving dynamic dataset, it has to be updated as some features of the training dataset are changed. This paper proposes a solution how to update this profile efficiently. Therefore, we investigate how to update the training profile when the data is constantly evolving. We assume that the data is modeled by a kernel method and processed by a spectral decomposition. In many algorithms for clustering and classification, a low dimensional representation of the affinity (kernel) graph of the embedded training dataset is computed. Then, it is used for classifying newly arrived data points. We present methods for updating such embeddings of the training datasets in an incremental way without the need to perform the entire computation upon the occurrences of changes in a small number of the training samples. Efficient computation of such an algorithm is critical in many web based applications.

Original languageEnglish
Pages (from-to)1356-1365
Number of pages10
JournalLinear Algebra and Its Applications
Issue number6
StatePublished - 15 Sep 2012


  • Diffusion Maps
  • Dimensionality reduction
  • Eigenvalue problem
  • Perturbation theory


Dive into the research topics of 'Updating kernel methods in spectral decomposition by affinity perturbations'. Together they form a unique fingerprint.

Cite this