Abstract
A class of fast identification algorithms is introduced for Gaussian process (GP) models. The fundamental approach is to propose a new kernel function which leads to a covariance matrix with low rank, a property that is consequently exploited for computational efficiency for both model parameter estimation and model predictions. The objective of either maximizing the marginal likelihood or the Kullback-Leibler (K-L) divergence between the estimated output probability density function (pdf) and the true pdf has been used as respective cost functions. For each cost function, an efficient coordinate descent algorithm is proposed to estimate the kernel parameters using a one dimensional derivative free search, and noise variance using a fast gradient descent algorithm. Numerical examples are included to demonstrate the effectiveness of the new identification approaches.
Original language | English |
---|---|
Pages (from-to) | 25-31 |
Number of pages | 7 |
Journal | Neurocomputing |
Volume | 133 |
DOIs | |
Publication status | Published - Jun 2014 |