The Intuitions Behind Bayesian Optimization with Gaussian Processes
Bayesian Optimization adds a Bayesian methodology to the iterative optimizer paradigm by incorporating a prior model on the space of possible target functions. This article introduces the basic concepts and intuitions behind Bayesian Optimization with Gaussian Processes.
By Charles Brecque, Mind Foundry
In certain applications the objective function is expensive or difficult to evaluate. In these situations, a general approach consists in creating a simpler surrogate model of the objective function which is cheaper to evaluate and will be used instead to solve the optimization problem. Moreover, due to the high cost of evaluating the objective function, an iterative approach is often recommended. Iterative optimizers work by iteratively requesting evaluations of the function at a sequence of points in the domain. Bayesian Optimization adds a Bayesian methodology to the iterative optimizer paradigm by incorporating a prior model on the space of possible target functions. This article introduces the basic concepts and intuitions behind Bayesian Optimization with Gaussian Processes and introduces OPTaaS, an API for Bayesian Optimization.
Optimization
Optimization methods try to locate an input x* in a domain 𝒳 to a function fwhich maximizes (or minimizes) the value of this function over 𝒳:
The general Optimization framework
In practice, the function f represents the outcome of a process that is required to be optimized, such as the overall profitability of a trading strategy, quality control metrics on a factory production line, or the performance of a data science pipeline with many parameters and hyperparameters.
The input domain 𝒳 represents the valid parameter choices for the process needing optimization. These could be market predictors used in a trading strategy, quantities of raw materials used in a factory process, or the parameters of ML models in a data science pipeline. It is the description of the input domain,𝒳, together with the properties of the function, f, that characterize the optimization problem. The valid inputs to the process domain, 𝒳, may be discrete, continuous, constrained, or any combination of these. Similarly, the outcome function, f, may be convex, differentiable, multi modal, noisy, slowly changing, or have many other important properties.
In certain applications the objective function is expensive to evaluate (computationally or economically), difficult to evaluate (chemical experiments, oil drilling). In these situations, a general approach consists in creating a simpler surrogate model f ̂of the objective function f which is cheaper to evaluate and will be used instead to solve the optimization problem.
Moreover, due to the high cost of evaluating the objective function, an iterative approach is often recommended. Iterative optimizers work by iteratively requesting evaluations of the function fat a sequence of points in the domain x1, x2, . . . ∈ 𝒳. Through these evaluations the optimizer is able to build up a picture of the function f. For gradient descent algorithms this picture is local but for surrogate model approaches this picture is global. At any time, or at the end of a preallocated budget of function evaluations, the iterative optimizer will be able to state its best approximation to the true value of x*.
The surrogate model is trained using N known evaluations of f: F =(f1, f2,…,fN ) at XN =(x1,x2,…,xN). There are many approaches used for building the surrogate model such as polynomial interpolation, neural networks, support vector machines, random forests and Gaussian processes. At Mind Foundry, our method of choice is regression using Gaussian Processes.
Gaussian Processes
Gaussian Processes (GPs) provide a rich and flexible class of nonparametric statistical models over function spaces with domains that can be continuous, discrete, mixed, or even hierarchical in nature. Furthermore, the GP provides not just information about the likely value of f, but importantly also about the uncertainty around that value.
The idea behind Gaussian Process Regression is for a set of observed values FN at some points XN we assume that these values correspond to the realisation of a multivariate Gaussian Process with a prior distribution:
where KN is a NxN covariance matrix and its coefficients are expressed in terms of a correlation function (or kernel) Kmn =K(xm,xn,θ). The hyper parameters θ of the kernel are calibrated according to a maximum likelihood principle. KN is chosen to reflect a prior assumption of the function and therefore the choice of the kernel will have a significant impact on the correctness of the regression. An illustration of several covariance functions is given in Figure 2.
Through mathematical transformations and using the conditional probability rule it is possible to estimate the posterior distribution p( f N+1FN , XN+1 ) and express f ̂N+1as a function of KN and FN with an uncertainty. This allows us to construct a probabilistic surrogate from our observations as illustrated in Figure 1:
Bayesian Optimization
Bayesian Optimization is a class of iterative optimization methods that focuses on the general optimization setting, where a description of 𝒳 is available, but knowledge of the properties of f is limited. Bayesian Optimization methods are characterized by two features:
 the surrogate model f ̂, for the function f,
 and an acquisition function computed from the surrogate and used for guiding the selection of the next evaluation point
BO adds a Bayesian methodology to the iterative optimizer paradigm by incorporating a prior model on the space of possible target functions, f. By updating this model every time time a function evaluation is reported, a Bayesian optimization routine keeps a posterior model of the target function f. This posterior model is the surrogate f ̂ for the function f. The pseudo code for a Bayesian Optimization routine with a GP prior is:
Initialisation:
 Place a Gaussian process prior on f
 Observe f at n0 points according to an initial spacefilling experimental design.
 Set n at n0
While n ≤ N do:
 Update the posterior probability distribution on f using all available data
 Identify the maximiser xn of the acquisition function over 𝒳, where the acquisition function is calculated using the current posterior distribution
 Observe yn = f (xn)
 Increment n
End while
Return either the point evaluated with the largest f (x) or the point with the largest posterior mean.
An example of a standard acquisition function is the Expected Improvement Criterion (EI), which, for any given point in x ∈ 𝒳 is the expected improvement in the value of f at x over the best value of f yet seen, given that the function f at x is indeed higher than the best value of f yet seen: so if we are finding the maxima of f, EI can be written as:
E I (x) = 𝔼(max( f (x) − f *,0))
where f * is the maximum value of f seen so far.
Additional examples of acquisition functions are:
 Entropy search which consists in seeking to minimise the uncertainty we have in the location of the optimal value
 Upper confidence bound
 Expected loss criterion
Figure 3 illustrates the evolution of the surrogate and its interactions with the acquisition function as it improves its knowledge after each iteration of the underlying function it is trying to minimize.
Operationalizing BO with OPTaaS
OPTaaS is a general purpose Bayesian optimizer which provides optimal parameter configurations via webservices. It can handle any parameter type and does not need to know the underlying process, models, or data. It requires the client to specify the parameters and their domains, and to post back the accuracy scores for each of OPTaaS’ recommended parameter configurations. OPTaaS uses these scores to model the underlying system and search for optimal configurations faster.
Mind Foundry have implemented an ensemble of surrogate models and acquisition functions within OPTaaS that it will automatically select and configure based on the nature and number of parameters it is provided as described in the Figure 4. This selection is based on thorough scientific testing and research so that OPTaaS always makes the most appropriate choices. Moreover, Mind Foundry has the ability to design custom covariance functions for a client’s specific problem which will significantly increase the speed and accuracy of the optimization process. Most users of OPTaaS require the optimization of complex processes, which are expensive to run, and give limited feedback. For this reason, OPTaaS focuses its API on providing a simple iterative optimizer interface. However, if more information is available about the process being optimized, it can always be leveraged for faster convergence towards the optimum. Therefore, OPTaaS also supports the communication of information about the domain 𝒳, such as constraints on inputs, and about the evaluations of the function f, such as noise or gradients or partially complete evaluations. Furthermore, clients are often able to leverage local infrastructure to distribute the optimization search, and OPTaaS can also be requested for batches to evaluate together.
The optimization process is the following:
 OPTaaS recommends a configuration to the customer
 The customer evaluates the configuration on their machines
 The customer sends back a score (accuracy, Sharpe ratio, Return on investment,…)
 OPTaaS uses the score to update its surrogate model and the cycle repeats until the optimal configuration has been reached.
During the whole process, OPTaaS does not access the underlying data or models. More information about OPTaaS can be found at the end of this page.
Team and Resources
Mind Foundry is an Oxford University spinout founded by Professors Stephen Roberts and Michael Osborne who have 35 person years in data analytics. The Mind Foundry team is composed of over 30 world class Machine Learning researchers and elite software engineers, many former postdocs from the University of Oxford. Moreover, Mind Foundry has a privileged access to over 30 Oxford University Machine Learning PhDs through its spinout status. Mind Foundry is a portfolio company of the University of Oxford and its investors include Oxford Sciences Innovation, the Oxford Technology and Innovations Fund, the University of Oxford Innovation Fund and Parkwalk Advisors.
Documentation
Tutorials: https://tutorial.optaas.mindfoundry.ai
API documentation: https://optaas.mindfoundry.ai
Research
http://www.robots.ox.ac.uk/~mosb/projects/project/2009/01/01/bayesopt/
References
Osborne, M.A. (2010). Bayesian Gaussian Processes for Sequential Prediction, Optimisation and Quadrature (PhD thesis). PhD thesis, University of Oxford.
Demo: charles.brecque@mindfoundry.ai
Bio: Charles Brecque is the Product Manager at Mind Foundry for OPTaaS, a general purpose Bayesian Optimizer deployed via webservices. Mind Foundry is an Oxford University spinout founded by Professors Stephen Roberts and Michael Osborne who have over 35 person years experience in advanced data analytics.
Original. Reposted with permission.
Related:
 Unfolding Naive Bayes From Scratch
 Optimization 101 for Data Scientists
 Beginners Ask “How Many Hidden Layers/Neurons to Use in Artificial Neural Networks?”
Top Stories Past 30 Days  


