Optimization as Estimation with Gaussian Processes in Bandit Settings (Supplement)
نویسندگان
چکیده
In this supplement, we provide proofs for all theorems and lemmas in the main paper, more exhaustive experimental results and details on the experiments. 1 Proofs 1.1 Proofs from Section 2 Lemma 2.1. In any round t, the point selected by EST is the same as the point selected by a variant of GP-UCB with λ t = min x∈Xˆmt−µt−1(x) σt−1(x). Conversely, the candidate selected by GP-UCB is the same as the candidate selected by a variant of EST withˆm t = max x∈X µ t−1 (x) + λ t σ t−1 (x). Proof. We omit the subscripts t for simplicity. Let a be the point selected by GP-UCB, and b selected by EST. Without loss of generality, we assume a and b are unique. With λ = min x∈Xˆm−µ(x) σ(x) , GP-UCB chooses to evaluate a = arg max x∈X µ(x)+λσ(x) = arg min x∈Xˆm − µ(x) σ(x). This is becausê m = max x∈X µ(x) + λσ(x) = µ(a) + λσ(a).
منابع مشابه
Optimization as Estimation with Gaussian Processes in Bandit Settings
Recently, there has been rising interest in Bayesian optimization – the optimization of an unknown function with assumptions usually expressed by a Gaussian Process (GP) prior. We study an optimization strategy that directly uses an estimate of the argmax of the function. This strategy offers both practical and theoretical advantages: no tradeoff parameter needs to be selected, and, moreover, w...
متن کاملOn 2-armed Gaussian Bandits and Optimization
We explore the 2-armed bandit with Gaussian payoos as a theoretical model for optimization. We formulate the problem from a Bayesian perspective, and provide the optimal strategy for both 1 and 2 pulls. We present regions of parameter space where a greedy strategy is provably optimal. We also compare the greedy and optimal strategies to a genetic-algorithm-based strategy. In doing so we correct...
متن کاملBatched Gaussian Process Bandit Optimization via Determinantal Point Processes
Gaussian Process bandit optimization has emerged as a powerful tool for optimizing noisy black box functions. One example in machine learning is hyper-parameter optimization where each evaluation of the target function may require training a model which may involve days or even weeks of computation. Most methods for this so-called “Bayesian optimization” only allow sequential exploration of the...
متن کاملOnline combinatorial optimization with stochastic decision sets and adversarial losses
Most work on sequential learning assumes a fixed set of actions that are available all the time. However, in practice, actions can consist of picking subsets of readings from sensors that may break from time to time, road segments that can be blocked or goods that are out of stock. In this paper we study learning algorithms that are able to deal with stochastic availability of such unreliable c...
متن کاملBayesian Estimation of Shift Point in Shape Parameter of Inverse Gaussian Distribution Under Different Loss Functions
In this paper, a Bayesian approach is proposed for shift point detection in an inverse Gaussian distribution. In this study, the mean parameter of inverse Gaussian distribution is assumed to be constant and shift points in shape parameter is considered. First the posterior distribution of shape parameter is obtained. Then the Bayes estimators are derived under a class of priors and using variou...
متن کامل