Scalable Hyperparameter Optimization with Lazy Gaussian Processes

2019 IEEE/ACM Workshop on Machine Learning in High Performance Computing Environments (MLHPC)(2019)

引用 5|浏览1
暂无评分
摘要
Most machine learning methods require careful selection of hyper-parameters in order to train a high performing model with good generalization abilities. Hence, several automatic selection algorithms have been introduced to overcome tedious manual (try and error) tuning of these parameters. Due to its very high sample efficiency, Bayesian Optimization over a Gaussian Processes modeling of the parameter space has become the method of choice. Unfortunately, this approach suffers from a cubic compute complexity due to underlying Cholesky factorization, which makes it very hard to be scaled beyond a small number of sampling steps. In this paper, we present a novel, highly accurate approximation of the underlying Gaussian Process. Reducing its computational complexity from cubic to quadratic allows an efficient strong scaling of Bayesian Optimization while outperforming the previous approach regarding optimization accuracy. First experiments show speedups of a factor of 162 in single node and further speed up by a factor of 5 in a parallel environment.
更多
查看译文
关键词
lazy Gaussian Processes,machine learning methods,hyper-parameters,high performing model,generalization abilities,automatic selection algorithms,high sample efficiency,Bayesian Optimization,Gaussian Processes modeling,parameter space,cubic compute complexity,Cholesky factorization,sampling steps,highly accurate approximation,computational complexity,efficient strong scaling,optimization accuracy,scalable hyperparameter Optimization
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要