Bisection Method

Suppose , and suppose we want to find such that

For special s we can get an exact solution in closed form, but this is the exception rather than the rule… In many cases, we can’t even write in closed form and instead treat it as a kind of black box with unpredictable behavior.

To find a root in the general case, we will work with an algorithm that generates a sequence where as , minimizing

In practice, we’re looking to satisfy , where is a computable error bound

We want an algorithm that is,

  • Robust - Always gives an approximate solution
  • Accurate - Gets closer with more computation with a computable error bound
  • Efficient - Gets to an approximation in as few evaluations as possible

The bisection method is a good example algorithm (but is not used in practice). It works because of the Intermediate Value Theorem; if we have an interval with a positive and negative value then there must be a zero. We iteratively cut our search interval in half, basically like binary search, until the length of the interval is satisfactory for our error bound.

How do we prove this formally? We have a few useful properties to work with:

It doesn’t take much experimenting to come up with the bound
We’d like to prove this bound formally

Theorem: Suppose , isa continuous function and . Then the sequence defined by the bisection algorithm converges to a limit such that , and .

To prove this, note that is bounded above and monotonically increasing and is bounded below and monotonically decreasing

and

So let’s define
The Pinching Theorem gives us

Since is continuous, we can use the inequalities to show that ,


So

To get the bound,



Simple Iteration

Another way to find is by writing and finding a fixed point

Theorem: Suppose is continuous and a contraction ( for some ), then has a unique fixed point . Furthermore, the sequence defined by converges to the unique fixed as , with

This method is robust, accurate, and efficient

Sometimes the conditions of our contraction mapping theorem are hard to check and we’d still like to know if convergence is possible

Theorem: Suppose is a continuous function, let be a fixed point of and assume that has a continuous derivative in some neighborhood of with , then converges to as , provided that is sufficiently close to

Newton’s Method

Newton’s method is what we use in practice, because of its convergence rate

Where does this equation come from? You can gain an intuition of it geometrically

Consider where is unknown and (i.e. )
We’d like to pick such that , yielding super-linear convergence

So we hope that in a neighborhood of we can get away with using

Theorem: Assume , , , for all , and where . Then as , where , and the convergence is at least quadratic, and precisely quadratic if .

How do we make sense of it? Let’s show the proof


is true by assumption, so we need to show


We use Taylor’s Theorem,
for some