The stochastic Ravine accelerated gradient method with general extrapolation coefficients - GREYC image Access content directly
Preprints, Working Papers, ... Year : 2024

The stochastic Ravine accelerated gradient method with general extrapolation coefficients

Abstract

In a real Hilbert space domain setting, we study the convergence properties of the stochastic Ravine accelerated gradient method for convex differentiable optimization. We consider the general form of this algorithm where the extrapolation coefficients can vary with each itera- tion, and where the evaluation of the gradient is subject to random errors. This general treatment models a breadth of practical algorithms and numerical implementations. We show that, under a proper tuning of the extrapolation parameters, and when the error variance associated with the gradient evaluations or the step-size sequences vanish sufficiently fast, the Ravine method provides fast convergence of the values both in expectation and almost surely. We also improve the con- vergence rates from O(·) to o(·) in expectation and almost sure sense. Moreover, we show almost sure summability property of the gradients, which implies the fast convergence of the gradients towards zero. This property reflects the fact that the high-resolution ODE of the Ravine method includes a Hessian-driven damping term. When the space is also separable, our analysis allows also to establish almost sure weak convergence of the sequence of iterates provided by the algorithm. We finally specialize the analysis to consider different parameter choices, including vanishing and constant (heavy ball method with friction) damping parameter, and present a comprehensive land- scape of the tradeoffs in speed and accuracy associated with these parameter choices and statistical properties on the sequence of errors in the gradient computations. We provide a thorough discus- sion of the similarities and differences with the Nesterov accelerated gradient which satisfies similar asymptotic convergence rates.
Fichier principal
Vignette du fichier
Ravine-general-coef-pert-preprint-submit-Mar-7-2024.pdf (415.4 Ko) Télécharger le fichier
Origin : Files produced by the author(s)

Dates and versions

hal-04506457 , version 1 (15-03-2024)
hal-04506457 , version 2 (19-03-2024)
hal-04506457 , version 3 (22-03-2024)

Identifiers

  • HAL Id : hal-04506457 , version 1

Cite

Hedy Attouch, Jalal M. Fadili, Vyacheslav Kungurtsev. The stochastic Ravine accelerated gradient method with general extrapolation coefficients. 2024. ⟨hal-04506457v1⟩
6 View
5 Download

Share

Gmail Facebook X LinkedIn More