BOHB: Robust and Efficient Hyperparameter Optimization at Scale

BOHB: Robust and Efficient Hyperparameter Optimization at Scale

4 Jul 2018 | Stefan Falkner, Aaron Klein, Frank Hutter
The paper "BOHB: Robust and Efficient Hyperparameter Optimization at Scale" by Stefan Falkner, Aaron Klein, and Frank Hutter addresses the challenges of hyperparameter optimization (HPO) in modern deep learning models, which are computationally expensive and sensitive to hyperparameters. The authors propose a new method, BOHB (Bayesian Optimization with Hyperband), that combines the strengths of Bayesian optimization and Hyperband to achieve strong anytime performance and fast convergence to optimal configurations. BOHB is designed to be robust, versatile, simple, and easy to implement, while also being scalable and efficient in parallel computing environments. The method is evaluated on a wide range of tasks, including high-dimensional toy functions, support vector machines, feed-forward neural networks, Bayesian neural networks, deep reinforcement learning, and convolutional neural networks. Empirical results demonstrate that BOHB consistently outperforms both Bayesian optimization and Hyperband, achieving both strong anytime performance and strong final performance. The paper also includes a detailed discussion of related work, an analysis of the method's hyperparameters, and a comparison to other combinations of Bayesian optimization and Hyperband.The paper "BOHB: Robust and Efficient Hyperparameter Optimization at Scale" by Stefan Falkner, Aaron Klein, and Frank Hutter addresses the challenges of hyperparameter optimization (HPO) in modern deep learning models, which are computationally expensive and sensitive to hyperparameters. The authors propose a new method, BOHB (Bayesian Optimization with Hyperband), that combines the strengths of Bayesian optimization and Hyperband to achieve strong anytime performance and fast convergence to optimal configurations. BOHB is designed to be robust, versatile, simple, and easy to implement, while also being scalable and efficient in parallel computing environments. The method is evaluated on a wide range of tasks, including high-dimensional toy functions, support vector machines, feed-forward neural networks, Bayesian neural networks, deep reinforcement learning, and convolutional neural networks. Empirical results demonstrate that BOHB consistently outperforms both Bayesian optimization and Hyperband, achieving both strong anytime performance and strong final performance. The paper also includes a detailed discussion of related work, an analysis of the method's hyperparameters, and a comparison to other combinations of Bayesian optimization and Hyperband.
Reach us at info@study.space