Is it possible to perform an

approximated fully Bayesian(1) selection of hyper-parameters (e.g. covariance scale) with the GPML code, instead of maximizing the marginal likelihood (2) ?

I think usingMCMC methodsto solve the integrals involving hyper-parameters prior should lead to better results when dealing with overfitting.

Up to my knowledge, the GPML framework doesn’t include these computations, but perhaps there are goods other third party codes.

(1) Sec. 5.2, Ch. 5 in Gaussian Process for Machine Learning, Rasmussen & Williams, 2006

**Answer**

There is another package for machine learning using Gaussian processes called GPstuff which has it all in my opinion. You can use MCMC, integration on a grid, etc. to marginalise out your hyperparameters.

NB In the documentation they call hyperparameters merely parameters.

**Attribution***Source : Link , Question Author : Emile , Answer Author : Mehrdad*