论文标题
用于一般凸功能的稳定高阶调谐器
A Stable High-order Tuner for General Convex Functions
论文作者
论文摘要
基于迭代梯度的算法已越来越多地用于培训包括大型神经网络在内的各种机器学习模型。尤其是,基于势头的方法具有加速的学习保证,由于其在某些类别的问题和多种算法中可证明可以保证快速学习,因此受到了很多关注。但是,这些方法的属性仅适用于恒定回归器。当发生时变回归器(在动态系统中很普遍)时,基于动量的方法中的许多方法无法保证稳定性。最近,为线性回归问题开发了一种新的高阶调谐器(HT),并显示为1)时间变化回归器的稳定性和渐近收敛性,以及2)非反应性加速学习保证的恒定回归。在本文中,我们扩展并讨论了同一HT的一般凸损失函数的结果。通过剥削凸度和平滑度定义,我们建立了相似的稳定性和渐近收敛的保证。最后,我们提供了支持HT算法以及加速学习属性令人满意行为的数值模拟。
Iterative gradient-based algorithms have been increasingly applied for the training of a broad variety of machine learning models including large neural-nets. In particular, momentum-based methods, with accelerated learning guarantees, have received a lot of attention due to their provable guarantees of fast learning in certain classes of problems and multiple algorithms have been derived. However, properties for these methods hold only for constant regressors. When time-varying regressors occur, which is commonplace in dynamic systems, many of these momentum-based methods cannot guarantee stability. Recently, a new High-order Tuner (HT) was developed for linear regression problems and shown to have 1) stability and asymptotic convergence for time-varying regressors and 2) non-asymptotic accelerated learning guarantees for constant regressors. In this paper, we extend and discuss the results of this same HT for general convex loss functions. Through the exploitation of convexity and smoothness definitions, we establish similar stability and asymptotic convergence guarantees. Finally, we provide numerical simulations supporting the satisfactory behavior of the HT algorithm as well as an accelerated learning property.