Hybrid Parameter Search and Dynamic Model Selection for Mixed-Variable Bayesian Optimization (2206.01409v4)
Abstract: This paper presents a new type of hybrid model for Bayesian optimization (BO) adept at managing mixed variables, encompassing both quantitative (continuous and integer) and qualitative (categorical) types. Our proposed new hybrid models (named hybridM) merge the Monte Carlo Tree Search structure (MCTS) for categorical variables with Gaussian Processes (GP) for continuous ones. hybridM leverages the upper confidence bound tree search (UCTS) for MCTS strategy, showcasing the tree architecture's integration into Bayesian optimization. Our innovations, including dynamic online kernel selection in the surrogate modeling phase and a unique UCTS search strategy, position our hybrid models as an advancement in mixed-variable surrogate models. Numerical experiments underscore the superiority of hybrid models, highlighting their potential in Bayesian optimization.
- Automatic kernel selection for gaussian processes regression with approximate bayesian computation and sequential monte carlo. Frontiers in Built Environment 3, 52.
- Auer, P. (2002). Using Confidence Bounds for Exploitation-Exploration Trade-Offs. Journal of Machine Learning Research 3, 397–422.
- Autotuning in high-performance computing applications. Proceedings of the IEEE 106(11), 2068–2083.
- Algorithms for Hyper-Parameter Optimization. Advances in neural information processing systems 24.
- Optimization by Gradient Boosting. In Advances in Contemporary Statistics and Econometrics, pp. 23–44. Springer.
- Structural kernel search via bayesian optimization and symbolical optimal transport. Advances in Neural Information Processing Systems 35, 39047–39058.
- Black-box Combinatorial Optimization using Models with Integer-valued Minima. Annals of Mathematics and Artificial Intelligence 89(7), 639–653.
- Similarity Encoding for Learning with Dirty Categorical Variables. Machine Learning 107(8), 1477–1494.
- BART: Bayesian Additive Regression Trees. The Annals of Applied Statistics 4(1), 266–298.
- Kernel methods for deep learning. Advances in neural information processing systems 22.
- Additive Gaussian Process for Computer Models With Qualitative and Quantitative Factors. Technometrics 59(3), 283–292.
- Bayesian Optimization over Hybrid Spaces. arXiv preprint arXiv:2106.04682.
- Neural architecture search: A survey. The Journal of Machine Learning Research 20(1), 1997–2017.
- Scaling multi-armed bandit algorithms. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 1449–1459.
- Friedman, J. H. (1991). Multivariate Adaptive Regression Splines. The Annals of Statistics, 1–67.
- Dealing with Categorical and Integer-valued Variables in Bayesian Optimization with Gaussian Processes . Neurocomputing 380, 20–35.
- STRUMPACK.
- An Efficient Multicore Implementation of a Novel HSS-structured Multifrontal Solver using Randomized Sampling. SIAM Journal on Scientific Computing 38(5), S358–S384.
- Algorithmic Assurance: An Active Approach to Algorithmic Testing Using Bayesian Optimisation. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, pp. 5470–5478.
- Gramacy, R. B. (2020). Surrogates: Gaussian Process Modeling, Design, and Optimization for the Applied Sciences.
- Hannan, E. J. and B. G. Quinn (1979). The Determination of the Order of an Autoregression. Journal of the Royal Statistical Society: Series B (Methodological) 41(2), 190–195.
- Harrison Jr, D. and D. L. Rubinfeld (1978). Hedonic Housing Prices and the Demand for Clean Air. Journal of environmental economics and management 5(1), 81–102.
- Scikit-Optimize/Scikit-Optimize. Zenodo.
- Recurrent kernel machines: Computing with infinite echo state networks. Neural Computation 24(1), 104–133.
- Sequential Model-Based Optimization for General Algorithm Configuration. In International Conference on Learning and Intelligent Optimization, pp. 507–523. Springer.
- Efficient global optimization of expensive black-box functions. Journal of Global optimization 13, 455–492.
- Continuous Surrogate-based Optimization Algorithms are Well-suited for Expensive Discrete Problems. In Benelux Conference on Artificial Intelligence, pp. 48–63. Springer.
- Bandit Based Monte-Carlo Planning. In D. Hutchison, T. Kanade, J. Kittler, J. M. Kleinberg, F. Mattern, J. C. Mitchell, M. Naor, O. Nierstrasz, C. Pandu Rangan, B. Steffen, M. Sudan, D. Terzopoulos, D. Tygar, M. Y. Vardi, G. Weikum, J. Fürnkranz, T. Scheffer, and M. Spiliopoulou (Eds.), Machine Learning: ECML 2006, Volume 4212, pp. 282–293. Berlin, Heidelberg: Springer Berlin Heidelberg.
- Efficient and robust bayesian selection of hyperparameters in dimension reduction for visualization. arXiv preprint arXiv:2306.00357.
- Non-smooth Bayesian Optimization in Tuning Problems. arXiv preprint arXiv:2109.07563.
- Luo, H. and M. T. Pratola (2022). Sharded Bayesian Additive Regression Trees. arXiv:2306.00361, 1–46.
- Optimism and Model Complexity Measure for Linear Models. In preparation.
- Bandits with adversarial scaling. In International Conference on Machine Learning, pp. 6511–6521. PMLR.
- Bayesian Optimization for Automated Model Selection. In D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett (Eds.), Advances in neural information processing systems, Volume 29. Curran Associates, Inc.
- Bayesian Optimization for Categorical and Category-Specific Continuous Inputs. arXiv:1911.12473 [cs, stat].
- Mixed Variable Bayesian Optimization with Frequency Modulated Kernels. arXiv preprint arXiv:2102.12792.
- Olson, R. S. and J. H. Moore (2019). TPOT: A Tree-Based Pipeline Optimization Tool for Automating Machine Learning. In F. Hutter, L. Kotthoff, and J. Vanschoren (Eds.), Automated Machine Learning, pp. 151–160. Cham: Springer International Publishing.
- Automated Machine Learning with Monte-Carlo Tree Search. arXiv:1906.00170 [cs, stat].
- Rasmussen, C. E. and C. K. I. Williams (2006). Gaussian Processes for Machine Learning. Adaptive Computation and Machine Learning. Cambridge, Mass: MIT Press.
- Bayesian Optimisation over Multiple Continuous and Categorical Inputs. arXiv:1906.08878 [cs, stat].
- Sethuraman, J. and R. C. Tiwari (1982). Convergence of Dirichlet Measures and the Interpretation of Their Parameter. In Statistical Decision Theory and Related Topics III, pp. 305–315. Elsevier.
- Taking the Human Out of the Loop: A Review of Bayesian Optimization. Proceedings of the IEEE 104(1), 148–175.
- Multitask and Transfer Learning for Autotuning Exascale Applications. arXiv:1908.05792 [cs, stat].
- GPTune User Guide.
- 3d-hst wfc3-selected photometric catalogs in the five candels/3d-hst fields: photometry, photometric redshifts, and stellar masses. The Astrophysical Journal Supplement Series 214(2), 24.
- Practical Bayesian Optimization of Machine Learning Algorithms. arXiv:1206.2944 [cs, stat].
- Virtual library of simulation experiments: Test functions and datasets. Retrieved June 2, 2022, from http://www.sfu.ca/~ssurjano.
- Raiders of the lost architecture: Kernels for bayesian optimization in conditional parameter spaces. arXiv preprint arXiv:1409.4011.
- Bayesian Inference in Monte-Carlo Tree Search. arXiv preprint arXiv:1203.3519.
- Bayesian Optimization for Auto-Tuning GPU kernels. arXiv:2111.14991 [cs, math].
- Ezgp: Easy-to-interpret gaussian process models for computer experiments with both quantitative and qualitative factors. SIAM/ASA Journal on Uncertainty Quantification 9(2), 333–353.
- Ye, J. (1998). On measuring and correcting the effects of data mining and model selection. Journal of the American Statistical Association 93(441), 120–131.
- Progressive sampling-based bayesian optimization for efficient and automatic machine learning model selection. Health information science and systems 5, 1–21.
- Bayesian optimization for materials design with mixed quantitative and qualitative variables. Scientific reports 10(1), 4924.
- A latent variable approach to Gaussian process modeling with qualitative and quantitative factors. Technometrics 62(3), 291–302.