Newton’s method

Newton’s method, sometimes called the Newton-Rhapson method, is a surprisingly simple and effective method for finding solutions to an equation of the form:

where is a differentiable function

Newton’s method is an iterative method, which means that we pick some starting point , and we follow some procedure to find , which is closer to the real root that satisfies . Then, we apply this method again and again, until we find some that approximates the real root with satisfactory accuracy.

I will illustrate the method on a function real-valued function . First, we pick a starting point . We then compute the linearized approximation to at this point (this is just the first-order Taylor expansion at ):

Now, we find the root of this approximation by solving for in

and we find . So implies . So, we have a new, supposedly better, approximation of the real root . Applying this iteratively gives:

which is the basic scheme behind Newton’s method.

We haven’t proved anything about the convergence of this method, and indeed, in general, the method might diverge. Note that it’s also possible to use an approximation of higher order, e.g.

However, using this method requires solving a more complicated solution. Instead of a simple linear equation we end up with a quadratic one.

Further, this method also works for multi-valued functions , with the restriction that should be an function. We can simply replace the factor by :

Convergence analysis

I will only consider the case where . I believe the general case where is messier, but the idea is essentially the same.

Take to be the root of , e.g. . Suppose that we have an estimate . We then have

Using the Taylor expansion of at , , and asymptotic notation gives:

From the first-order Taylor expansion of at we find

From the above two equalities we find

So, if is sufficiently small, each iteration of Newton’s method reduces the error to the square of the error of the previous iteration. We say that Newton’s method has quadratic convergence.

In practice, this is also observed:

| | | | |--|--|--| | 0 | 1 | 0.414213562373095 | | 1 | 1.5 | 0.085786437626905 | | 2 | 1.416666666666667 | 0.002453104293572 | | 3 | 1.414215686274510 | 0.000002123901415 | | 4 | 1.414213562374690 | 0.000000000001595 |

The number of leading zeros in the error roughly doubles each iteration, leading to 11 correct decimal digits after just four iterations.

Picking initial values

One problem with Newton’s method is that it’s sensitive to the initial value. Sometimes it’s possible to make an educated guess, but often not. If this is a problem, one might consider using a homotopy method. This description is based on the description given in the book "Numerical Methods in Scientific Computing", by J. van Kan, A. Segal, and F. Vermolen. In this method, one picks an equation with a known solution , and considers the problem

And now one increases in small steps. Every time, Newton’s method is used to find the solution of . The solution found this way is then used as an initial estimate for the next step.