Incremental sparse Pseudo-input Gaussian process regression

Research output: Contribution to journalArticle

1 Citation (Scopus)

Abstract

In this paper, we devise a novel method that incrementally learns pseudo-data, which represent the whole training data set for Gaussian Process (GP) regression. The method involves sparse approximation of the GP by extending the work of Snelson and Ghahramani. We call the proposed method Incremental Sparse Pseudo-input Gaussian Process (ISPGP) regression. Unlike the Snelson and Ghahramani's work, the proposed ISPGP algorithm allows for training from either a huge amount of training data by scanning through it only once or an online incremental training data set. We also design a likelihood weighting scheme to incrementally determine pseudo-data while maintaining the representational power. Due to the nature of the incremental learning algorithm, the proposed ISPGP algorithm can theoretically work with infinite data to which the conventional GP or Sparse Pseudo-input Gaussian Process (SPGP) algorithm is not applicable. From our experimental results on the KIN40K data set, we can see that the proposed ISPGP algorithm is comparable to the conventional GP algorithm using the same number of training data. It also significantly reduces the computational cost and memory requirement in regression and is scalable to a large training data set without significant performance degradation. Although the proposed ISPGP algorithm performs slightly worse than Snelson and Ghahramani's SPGP algorithm, the level of performance degradation is acceptable.

Original languageEnglish
Article number1250019
JournalInternational Journal of Pattern Recognition and Artificial Intelligence
Volume26
Issue number8
DOIs
Publication statusPublished - 2012 Dec 1

    Fingerprint

Keywords

  • Gaussian process regression
  • incremental learning
  • pseudo-data

ASJC Scopus subject areas

  • Artificial Intelligence
  • Software
  • Computer Vision and Pattern Recognition

Cite this