Description: Conjugate Gradient method (Conjugate Gradient) is between the steepest descent method between Newton method and a method, it only USES a derivative information, but overcome the steepest descent method slow convergence of weakness, but also avoid the Newton law needs to storage and computing Hesse inverse matrix and shortcomings, Conjugate Gradient method is not only solve linear equations with most of the large method, and also one of the most effective solution large nonlinear optimization of one of the algorithm. In all kinds of optimization algorithm, the conjugate gradient method is very important. Its advantage is the storage capacity needed, it has small step convergence, high stability, and doesn t require any exotic parameters.
To Search:
File list (Check if you may need any files):
Newton.m