On some Curvature-Dependent Steplength for the Gradient Method
Journal of convex analysis, Tome 17 (2010) no. 3, pp. 765-78
The aim of this paper is to show the interest of taking into account the notion of curvature in gradient methods. More precisely, given a Hilbert space $H$ and a strictly convex function $\phi:H\to {\mathbb{R}}$ of class ${\mathcal C}^2$, we consider the following algorithm $$x_{n+1}=x_n-\lambda_n\, \nabla \phi(x_n),\quad \mbox{ with } \lambda_n = \frac{|\nabla \phi(x_n)|^2}{\langle\nabla^2\phi(x_n).\nabla\phi(x_n), \nabla\phi(x_n)\rangle}.\leqno (\star)$$ We obtain results of linear convergence for the above algorithm, even without strong convexity. Some variants of $(\star)$ are also considered, with different expressions of the curvature-dependent steplength $\lambda_n$. A large part of the paper is devoted to the study of an implicit version of $(\star)$, falling into the field of the proximal point iteration. All these algorithms are clearly related to the Barzilai-Borwein method and numerical illustrations at the end of the paper allow to compare these different schemes.
Classification :
65K10, 90C25, 49M25
Mots-clés : Unconstrained convex optimization, steepest descent, gradient method, proximal point algorithm, Barzilai-Borwein stepsize
Mots-clés : Unconstrained convex optimization, steepest descent, gradient method, proximal point algorithm, Barzilai-Borwein stepsize
@article{JCA_2010_17_3_JCA_2010_17_3_a4,
author = {B. Baji and A. Cabot},
title = {On some {Curvature-Dependent} {Steplength} for the {Gradient} {Method}},
journal = {Journal of convex analysis},
pages = {765--78},
year = {2010},
volume = {17},
number = {3},
url = {http://geodesic.mathdoc.fr/item/JCA_2010_17_3_JCA_2010_17_3_a4/}
}
B. Baji; A. Cabot. On some Curvature-Dependent Steplength for the Gradient Method. Journal of convex analysis, Tome 17 (2010) no. 3, pp. 765-78. http://geodesic.mathdoc.fr/item/JCA_2010_17_3_JCA_2010_17_3_a4/