EigenGP: Sparse Gaussian process models with data-dependent eigenfunctions
Abstract
Gaussian processes (GPs) provide a nonparametric representation of functions. However, classical GP inference suffers from high computational cost and it is difficult to design nonstationary GP priors in practice. In this paper, we propose a sparse Gaussian process model, EigenGP, based on the Karhunen-Loeve (KL) expansion of a GP prior. We use the Nystrom approximation to obtain data dependent eigenfunctions and select these eigenfunctions by evidence maximization. This selection reduces the number of eigenfunctions in our model and provides a nonstationary covariance function. To handle nonlinear likelihoods, we develop an efficient expectation propagation (EP) inference algorithm, and couple it with expectation maximization for eigenfunction selection. Because the eigenfunctions of a Gaussian kernel are associated with clusters of samples - including both the labeled and unlabeled - selecting relevant eigenfunctions enables EigenGP to conduct semi-supervised learning. Our experimental results demonstrate improved predictive performance of EigenGP over alternative state-of-the-art sparse GP and semisupervised learning methods for regression, classification, and semisupervised classification.
- Publication:
-
arXiv e-prints
- Pub Date:
- April 2012
- DOI:
- 10.48550/arXiv.1204.3972
- arXiv:
- arXiv:1204.3972
- Bibcode:
- 2012arXiv1204.3972Q
- Keywords:
-
- Computer Science - Machine Learning;
- Statistics - Computation;
- Statistics - Machine Learning
- E-Print:
- 10 pages, 19 figures