An accelerated minimal gradient method with momentum for strictly convex quadratic optimization

Harry Oviedo, Oscar Dalmau, Rafael Herrera

Research output: Contribution to journalArticlepeer-review

1 Scopus citations


In this article we address the problem of minimizing a strictly convex quadratic function using a novel iterative method. The new algorithm is based on the well-known Nesterov’s accelerated gradient method. At each iteration of our scheme, the new point is computed by performing a line-search scheme using a search direction given by a linear combination of three terms, whose parameters are chosen so that the residual norm is minimized at each step of the process. We establish the linear convergence of the proposed method and show that its convergence rate factor is analogous to the one available for other gradient methods. Finally, we present preliminary numerical results on some sets of synthetic and real strictly convex quadratic problems, showing that the proposed method outperforms in terms of efficiency, a wide collection of state-of-the art gradient methods, and that it is competitive against the conjugate gradient method in terms of CPU time and number of iterations.

Original languageEnglish
Pages (from-to)591-606
Number of pages16
JournalBIT Numerical Mathematics
Issue number2
StatePublished - Jun 2022
Externally publishedYes


  • Accelerated gradient methods
  • Convex optimization
  • Gradient method
  • Linear system of equations


Dive into the research topics of 'An accelerated minimal gradient method with momentum for strictly convex quadratic optimization'. Together they form a unique fingerprint.

Cite this