Learning to Warm-Start Bayesian Hyperparameter Optimization

10/17/2017
by   Saehoon Kim, et al.
0

Hyperparameter optimization undergoes extensive evaluations of validation errors in order to find its best configuration. Bayesian optimization is now popular for hyperparameter optimization, since it reduces the number of validation error evaluations required. Suppose that we are given a collection of datasets on which hyperparameters are already tuned by either humans with domain expertise or extensive trials of cross-validation. When a model is applied to a new dataset, it is desirable to let Bayesian optimization start from configurations that were successful on similar datasets. To this end, we construct a Siamese network with convolutional layers followed by bi-directional LSTM layers, to learn meta-features over image datasets. Learned meta-features are used to select a few datasets that are similar to the new dataset, so that a set of configurations in similar datasets is adopted as initialization to warm-start Bayesian hyperparameter optimization. Experiments on image datasets demonstrate that our learned meta-features are useful in optimizing hyperparameters in deep residual networks for image classification.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset