Bisection Method
Suppose , and suppose we want to find such that
For special s we can get an exact solution in closed form, but this is the exception rather than the rule… In many cases, we can’t even write in closed form and instead treat it as a kind of black box with unpredictable behavior.
To find a root in the general case, we will work with an algorithm that generates a sequence where as , minimizing
In practice, we’re looking to satisfy , where is a computable error bound
We want an algorithm that is,
- Robust - Always gives an approximate solution
- Accurate - Gets closer with more computation with a computable error bound
- Efficient - Gets to an approximation in as few evaluations as possible
The bisection method is a good example algorithm (but is not used in practice). It works because of the Intermediate Value Theorem; if we have an interval with a positive and negative value then there must be a zero. We iteratively cut our search interval in half, basically like binary search, until the length of the interval is satisfactory for our error bound.
How do we prove this formally? We have a few useful properties to work with:
It doesn’t take much experimenting to come up with the bound
We’d like to prove this bound formally
Theorem: Suppose , isa continuous function and . Then the sequence defined by the bisection algorithm converges to a limit such that , and .
To prove this, note that is bounded above and monotonically increasing and is bounded below and monotonically decreasing
and
So let’s define
The Pinching Theorem gives us
Since is continuous, we can use the inequalities to show that ,
So
To get the bound,
Simple Iteration
Another way to find is by writing and finding a fixed point
Theorem: Suppose is continuous and a contraction ( for some ), then has a unique fixed point . Furthermore, the sequence defined by converges to the unique fixed as , with
This method is robust, accurate, and efficient
Sometimes the conditions of our contraction mapping theorem are hard to check and we’d still like to know if convergence is possible
Theorem: Suppose is a continuous function, let be a fixed point of and assume that has a continuous derivative in some neighborhood of with , then converges to as , provided that is sufficiently close to
Newton’s Method
Newton’s method is what we use in practice, because of its convergence rate
Where does this equation come from? You can gain an intuition of it geometrically
Consider where is unknown and (i.e. )
We’d like to pick such that , yielding super-linear convergence
So we hope that in a neighborhood of we can get away with using
Theorem: Assume , , , for all , and where . Then as , where , and the convergence is at least quadratic, and precisely quadratic if .
How do we make sense of it? Let’s show the proof
is true by assumption, so we need to show
We use Taylor’s Theorem,
for some