An Advanced Conjugate Gradient Training Algorithm Based on a Modified Secant Equation

Ioannis E. Livieris, Panagiotis Pintelas
2012 ISRN Artificial Intelligence  
Conjugate gradient methods constitute excellent neural network training methods characterized by their simplicity, numerical efficiency, and their very low memory requirements. In this paper, we propose a conjugate gradient neural network training algorithm which guarantees sufficient descent using any line search, avoiding thereby the usually inefficient restarts. Moreover, it achieves a high-order accuracy in approximating the second-order curvature information of the error surface by
more » ... surface by utilizing the modified secant condition proposed by Li et al. (2007). Under mild conditions, we establish that the proposed method is globally convergent for general functions under the strong Wolfe conditions. Experimental results provide evidence that our proposed method is preferable and in general superior to the classical conjugate gradient methods and has a potential to significantly enhance the computational efficiency and robustness of the training process.
doi:10.5402/2012/486361 fatcat:fsaksmihpjhpdhxxtbnfocnbyq