论文标题
学会成为全球优化器
Learning to be Global Optimizer
论文作者
论文摘要
人工智能的发展为优化算法的发展提供了新的启示。本文建议学习一个两相(包括最小化阶段和逃逸阶段)的全局优化算法,以实现平滑的非凸功能。对于最小化阶段,开发了一种模型驱动的深度学习方法,以了解下降方向的更新规则,该方法被形式化为历史信息的非线性组合,以供凸功能。我们证明,所提出的自适应方向的最终算法可以保证凸功能的收敛。实证研究表明,学到的算法显着超过了一些众所周知的经典优化算法,例如梯度下降,结合下降和BFG,并且在不足的功能上表现良好。从本地最佳限度开始逃脱阶段,以固定的逃脱策略为马尔可夫决策过程。我们进一步建议通过加强学习来学习最佳的逃避政策。通过优化合成功能并训练深层神经网络以进行CIFAR图像分类来验证逃脱策略的有效性。学到的两相全局优化算法展示了一些基准功能和机器学习任务的有希望的全局搜索能力。
The advancement of artificial intelligence has cast a new light on the development of optimization algorithm. This paper proposes to learn a two-phase (including a minimization phase and an escaping phase) global optimization algorithm for smooth non-convex functions. For the minimization phase, a model-driven deep learning method is developed to learn the update rule of descent direction, which is formalized as a nonlinear combination of historical information, for convex functions. We prove that the resultant algorithm with the proposed adaptive direction guarantees convergence for convex functions. Empirical study shows that the learned algorithm significantly outperforms some well-known classical optimization algorithms, such as gradient descent, conjugate descent and BFGS, and performs well on ill-posed functions. The escaping phase from local optimum is modeled as a Markov decision process with a fixed escaping policy. We further propose to learn an optimal escaping policy by reinforcement learning. The effectiveness of the escaping policies is verified by optimizing synthesized functions and training a deep neural network for CIFAR image classification. The learned two-phase global optimization algorithm demonstrates a promising global search capability on some benchmark functions and machine learning tasks.