BANANAS: Bayesian Optimization with Neural Architectures for Neural Architecture Search
نویسندگان
چکیده
Over the past half-decade, many methods have been considered for neural architecture search (NAS). Bayesian optimization (BO), which has long had success in hyperparameter optimization, recently emerged as a very promising strategy NAS when it is coupled with predictor. Recent work proposed different instantiations of this framework, example, using networks or graph convolutional predictive model within BO. However, analyses these papers often focus on full-fledged algorithm, so difficult to tell individual components framework lead best performance. In work, we give thorough analysis "BO + predictor framework" by identifying five main components: encoding, predictor, uncertainty calibration method, acquisition function, and function optimization. We test several each component also develop novel path-based encoding scheme architectures, show theoretically empirically scales better than other encodings. Using all our analyses, final algorithm called BANANAS, achieves state-of-the-art performance spaces. adhere research checklist (Lindauer Hutter 2019) facilitate practices, code available at https://github.com/naszilla/naszilla.
منابع مشابه
Neural Architecture Search with Bayesian Optimisation and Optimal Transport
Abstract Bayesian Optimisation (BO) refers to a class of methods for global optimisation of a function f which is only accessible via point evaluations. It is typically used in settings where f is expensive to evaluate. A common use case for BO in machine learning is model selection, where it is not possible to analytically model the generalisation performance of a statistical model, and we res...
متن کاملBayesian Optimization with Robust Bayesian Neural Networks
Bayesian optimization is a prominent method for optimizing expensive-to-evaluate black-box functions that is widely applied to tuning the hyperparameters of machine learning algorithms. Despite its successes, the prototypical Bayesian optimization approach – using Gaussian process models – does not scale well to either many hyperparameters or many function evaluations. Attacking this lack of sc...
متن کاملBayesian Optimization Combined with Successive Halving for Neural Network Architecture Optimization
The choice of hyperparameters and the selection of algorithms is a crucial part in machine learning. Bayesian optimization methods and successive halving have been applied successfully to optimize hyperparameters automatically. Therefore, we propose to combine both methods by estimating the initial population of incremental evaluation, our variation of successive halving, by means of Bayesian o...
متن کاملProgressive Neural Architecture Search
We propose a method for learning CNN structures that is more efficient than previous approaches: instead of using reinforcement learning (RL) or genetic algorithms (GA), we use a sequential model-based optimization (SMBO) strategy, in which we search for architectures in order of increasing complexity, while simultaneously learning a surrogate function to guide the search, similar to A* search....
متن کاملNeural Architecture Search with Reinforcement Learning
Neural networks are powerful and flexible models that work well for many difficult learning tasks in image, speech and natural language understanding. Despite their success, neural networks are still hard to design. In this paper, we use a recurrent network to generate the model descriptions of neural networks and train this RNN with reinforcement learning to maximize the expected accuracy of t...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2021
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v35i12.17233