Enhancing Evolutionary Optimization in Uncertain Environments by Allocating Evaluations via Multi-armed Bandit Algorithms

نویسندگان

  • Xin Qiu
  • Risto Miikkulainen
چکیده

Optimization problems with uncertain fitness functions are common in the real world, and present unique challenges for evolutionary optimization approaches. Existing issues include excessively expensive evaluation, lack of solution reliability, and incapability in maintaining high overall fitness during optimization. Using conversion rate optimization as an example, this paper proposes a series of new techniques for addressing these issues. The main innovation is to augment evolutionary algorithms by allocating evaluation budget through multi-armed bandit algorithms. Experimental results demonstrate that multi-armed bandit algorithms can be used to allocate evaluations efficiently, select the winning solution reliably and increase overall fitness during exploration. The proposed methods can be generalized to any optimization problems with noisy fitness functions.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Optimization in Uncertain and Complex Dynamic Environments with Evolutionary Methods

In the real world, many of the optimization issues are dynamic, uncertain, and complex in which the objective function or constraints can be changed over time. Consequently, the optimum of these issues is changed nonlinearly. Therefore, the optimization algorithms not only should search the global optimum value in the space but also should follow the path of optimal change in dynamic environmen...

متن کامل

Reflections on Bandit Problems and Selection Methods in Uncertain Environments

The behavior of selection methods used in evolutionary algorithms that operate in uncertain environments is investigated in the framework of parametric two{armed bandit problems. Asymptotically optimal selection strategies are based on the sequential probability ratio test which i s p r o ved to perform up to four times better than analogous strategies based on the optimal xed size sample test....

متن کامل

From Bandits to Monte-Carlo Tree Search: The Optimistic Principle Applied to Optimization and Planning

This work covers several aspects of the optimism in the face of uncertainty principle applied to large scale optimization problems under finite numerical budget. The initial motivation for the research reported here originated from the empirical success of the so-called Monte-Carlo Tree Search method popularized in Computer Go and further extended to many other games as well as optimization and...

متن کامل

Planning in Reward-Rich Domains via PAC Bandits

In some decision-making environments, successful solutions are common. If the evaluation of candidate solutions is noisy, however, the challenge is knowing when a “good enough” answer has been found. We formalize this problem as an infinite-armed bandit and provide upper and lower bounds on the number of evaluations or “pulls” needed to identify a solution whose evaluation exceeds a given thres...

متن کامل

The Max K-Armed Bandit: A New Model of Exploration Applied to Search Heuristic Selection

The multiarmed bandit is often used as an analogy for the tradeoff between exploration and exploitation in search problems. The classic problem involves allocating trials to the arms of a multiarmed slot machine to maximize the expected sum of rewards. We pose a new variation of the multiarmed bandit—the Max K-Armed Bandit—in which trials must be allocated among the arms to maximize the expecte...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2018