Using the full posterior distribution with the hyperparameters as unknown variables is known to give better results in Bayesian optimization (see https://arxiv.org/pdf/1206.2944.pdf).
A user could opt-in to using this technique by replacing the MAPGPOptimizer with MCMCEstimate or another appropriate name. GaussianProcesses provides an mcmc function to estimate hyperparameters but my understanding of the source is that it does not marginalize over the hyperparameters and compute an integrated acquisition function (which I suppose wouldn't make sense within the scope of GaussianProcesses).
Thoughts on including something like this? The way I see it, the work would break down as follows:
Using the full posterior distribution with the hyperparameters as unknown variables is known to give better results in Bayesian optimization (see https://arxiv.org/pdf/1206.2944.pdf).
A user could opt-in to using this technique by replacing the
MAPGPOptimizerwithMCMCEstimateor another appropriate name.GaussianProcessesprovides anmcmcfunction to estimate hyperparameters but my understanding of the source is that it does not marginalize over the hyperparameters and compute an integrated acquisition function (which I suppose wouldn't make sense within the scope ofGaussianProcesses).Thoughts on including something like this? The way I see it, the work would break down as follows:
MCMCEstimate