Gaussian process : function approximation properties

I’m learning about Gaussian Process and have heard only bits and pieces. Would really appreciate comments and answers.

For any set of data, is it true that a Gaussian Process function approximation would give zero or negligible fitting error at the data points ? In another place I also heard that Gaussian Process is particularly good for noisy data. This seems to be in conflict with the low fitting error for any observed data?

Additionally, further away from the data points there seem to be more uncertainty (larger covariance). If so, does it behave like local models (RBF etc)?

Finally, is there any universal approximation property?


Suppose data sample is D=(X,y)={xi,yi=y(xi)}Ni=1.
Also suppose, that we have a covariance function k(x1,x2) and zero mean specified for a Gussian process. Distribution for a new point x will be Gaussian with mean m(x)=kK1y and variance V(x)=k(x,x)kK1kT. Vector k={k(x,x1),,k(x,xN)} is a vector of covariances, matrix K={k(xi,xj)}Ni,j=1 is a matrix of sample covariances. In case we make prediction using mean value of posterior distribution for sample interpolation property holds. Really,
But, it isn’t the case if we use regularization i.e. incorporate white noise term. in this case covariance matrix for sample has form K+σI, but for covariances with real function values we have covariance matrix K, and posterior mean is
In addition, regularization makes problem more computationally stable.

Choosing noise variance σ we can select if we want interpolation (σ=0) or we want to handle noisy observations (σ is big).

Also, the Gaussian processes regression is local method because variance of predictions grows with distance to learning sample, but we can select appropriate covariance function k and handle more complex problems, than with RBF. Another nice property is small number of parameters. Usually it equals O(n), where n is data dimension.

Source : Link , Question Author : oalah , Answer Author : Alexey Zaytsev

Leave a Comment