首页|Convergence analysis of AdaBound with relaxed bound functions for non-convex optimization

Convergence analysis of AdaBound with relaxed bound functions for non-convex optimization

扫码查看
? 2021 Elsevier LtdClipping on learning rates in Adam leads to an effective stochastic algorithm—AdaBound. In spite of its effectiveness in practice, convergence analysis of AdaBound has not been fully explored, especially for non-convex optimization. To this end, we address the convergence of the last individual output of AdaBound for non-convex stochastic optimization problems, which is called individual convergence. We prove that, with the iteration of the AdaBound, the cost function converges to a finite value and the corresponding gradient converges to zero. The novelty of this proof is that the convergence conditions on the bound functions and momentum factors are much more relaxed than the existing results, especially when we remove the monotonicity and convergence of the bound functions, and only keep their boundedness. The momentum factors can be fixed to be constant, without the restriction of monotonically decreasing. This provides a new perspective on understanding the bound functions and momentum factors of AdaBound. At last, numerical experiments are provided to corroborate our theory and show that the convergence of AdaBound extends to more general bound functions.

AdaBoundBound functionsDeep learningIndividual convergenceNon-convex optimization

Liu J.、Kong J.、Xu D.、Qi M.、Lu Y.

展开 >

School of Mathematics and Statistics Northeast Normal University

Institute for Intelligent Elderly Care Changchun Humanities and Sciences College

2022

Neural Networks

Neural Networks

EISCI
ISSN:0893-6080
年,卷(期):2022.145
  • 6
  • 33