The gradient method is based on the first-order term in the Taylor
expansion for
. By taking a second-order term as well and
solving the quadratic minimization problem iteratively, Newton's
method for functional minimization is obtained. Essentially,
Newton's method requires the error surface to be close to
quadratic, and its effectiveness is directly tied to the
accuracy of this assumption. For most problems, the error surface can
be well approximated by a quadratic form near the solution. For this
reason, Newton's method tends to give very rapid (``quadratic'')
convergence in the last stages of iteration.
Newton's method is derived as follows: The Taylor expansion of
about
gives
Definition. Newton's method is defined by
When
, the
answer is obtained after the first iteration. In particular, when the
error surface
is a quadratic form in
, Newton's
method produces
in one iteration, i.e.,
for
every
.
For Newton's method, there is the following general result on the existence of a critical point (i.e., a point at which the gradient vanishes) within a sphere of a Banach space.
Theorem. (Kantorovich)
Let
be a point in
for which
exists, and set
Proof. See Goldstein [2], p. 143.