论文标题
贝叶斯优化用于选择高效的机器学习模型
Bayesian Optimization for Selecting Efficient Machine Learning Models
论文作者
论文摘要
许多机器学习模型的性能取决于其超参数设置。贝叶斯优化已成为机器学习算法高参数优化的成功工具,该算法旨在在迭代顺序过程中识别最佳的超参数。但是,大多数贝叶斯优化算法旨在仅选择有效性的模型,而忽略了模型训练效率的重要问题。鉴于模型有效性和训练时间对于现实世界的应用都很重要,因此选择有效性的模型可能无法满足在生产环境中部署所需的严格培训时间要求。在这项工作中,我们提出了一个统一的贝叶斯优化框架,用于共同优化预测有效性和训练效率的模型。我们提出了一个目标,该目标捕获了这两个指标之间的权衡,并证明了我们如何在有原则的贝叶斯优化框架中共同优化它们。针对建议任务的模型选择实验表明,与最先进的贝叶斯优化算法相比,选择这种模型可显着提高模型训练效率,同时保持强大的效率。
The performance of many machine learning models depends on their hyper-parameter settings. Bayesian Optimization has become a successful tool for hyper-parameter optimization of machine learning algorithms, which aims to identify optimal hyper-parameters during an iterative sequential process. However, most of the Bayesian Optimization algorithms are designed to select models for effectiveness only and ignore the important issue of model training efficiency. Given that both model effectiveness and training time are important for real-world applications, models selected for effectiveness may not meet the strict training time requirements necessary to deploy in a production environment. In this work, we present a unified Bayesian Optimization framework for jointly optimizing models for both prediction effectiveness and training efficiency. We propose an objective that captures the tradeoff between these two metrics and demonstrate how we can jointly optimize them in a principled Bayesian Optimization framework. Experiments on model selection for recommendation tasks indicate models selected this way significantly improves model training efficiency while maintaining strong effectiveness as compared to state-of-the-art Bayesian Optimization algorithms.