A New Class of Conjugate Gradient Methods Based on a Modified Search Direction for Unconstrained Optimization
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 84420
A New Class of Conjugate Gradient Methods Based on a Modified Search Direction for Unconstrained Optimization

Authors: Belloufi Mohammed, Sellami Badreddine

Abstract:

Conjugate gradient methods have played a special role for solving large scale optimization problems due to the simplicity of their iteration, convergence properties and their low memory requirements. In this work, we propose a new class of conjugate gradient methods which ensures sufficient descent. Moreover, we propose a new search direction with the Wolfe line search technique for solving unconstrained optimization problems, a global convergence result for general functions is established provided that the line search satisfies the Wolfe conditions. Our numerical experiments indicate that our proposed methods are preferable and in general superior to the classical conjugate gradient methods in terms of efficiency and robustness.

Keywords: unconstrained optimization, conjugate gradient method, sufficient descent property, numerical comparisons

Procedia PDF Downloads 366