Description: Conjugate gradient method is between the steepest descent method and Newton method between a method that only use the first derivative information, but the steepest descent method to overcome the disadvantage of slow convergence, but also avoids the need to store and calculate Newton Hesse matrix and the shortcomings of the inverse, conjugate gradient method is not only linear equations to solve large-scale one of the most useful, large-scale nonlinear optimization solution is also the most efficient algorithms.
To Search:
File list (Check if you may need any files):
共轭梯度法\conjgrad.m
..........\license.txt
共轭梯度法