Optimistic Optimization of a Deterministic Function without the Knowledge of its Smoothness
–Neural Information Processing Systems
We consider a global optimization problem of a deterministic function f in a semimetric space,given a finite budget ofnevaluations. The functionf is assumed to be locally smooth (around one of its global maxima) with respect to a semi-metric l. We describe two algorithms based on optimistic exploration that use a hierarchical partitioningof the space at all scales. A first contribution is an algorithm, DOO, that requires the knowledge of l. We report a finite-sample performance bound in terms of a measure of the quantity of near-optimal states. We then define a second algorithm, SOO, which does not require the knowledge of the semimetric lunder which f is smooth, and whose performance is almost as good as DOO optimally-fitted.
Neural Information Processing Systems
Dec-31-2011