Namespaces
Variants
Actions

Gradient method

From Encyclopedia of Mathematics
Jump to: navigation, search
The printable version is no longer supported and may have rendering errors. Please update your browser bookmarks and please use the default browser print function instead.


A method for the minimization of a function of several variables. It is based on the fact that each successive approximation of the function $ F $ is obtained from the preceding one by a shift in the direction of the gradient of the function:

$$ \mathbf x ^ {n + 1 } = \ \mathbf x ^ {n} - \delta _ {n} \ \mathop{\rm grad} F ( \mathbf x ^ {n} ). $$

The parameter $ \delta _ {n} $ can be obtained, e.g., from the condition of the magnitude

$$ F ( \mathbf x ^ {n} - \delta _ {n} \ \mathop{\rm grad} F ( \mathbf x ^ {n} )) \ \ \textrm{ being minimal } . $$

See also Descent, method of; Steepest descent, method of.

Comments

References

[a1] J.E. Dennis, R.B. Schnabel, "Numerical methods for unconstrained optimization and nonlinear equations" , Prentice-Hall (1983) MR0702023 Zbl 0579.65058
[a2] R. Fletcher, "Practical methods of optimization" , Wiley (1980) MR0585160 MR0633058 Zbl 0439.93001
[a3] D.G. Luenberger, "Linear and nonlinear programming" , Addison-Wesley (1984) MR2423726 MR2012832 Zbl 0571.90051
How to Cite This Entry:
Gradient method. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Gradient_method&oldid=47111