Newton's Method#
Newton's method is a numerical algorithm for finding solutions to equations of the form \(f(x) = 0\), where \(f: \mathcal{D} \subseteq \mathbb{R} \to \mathbb{R}\) is a real function, or to systems of equations which can be written as \(f(\boldsymbol{x}) = \boldsymbol{0}\) for some real vector field \(f: \mathcal{D} \subseteq \mathbb{R}^n \to \mathbb{R}^n\).
In One Dimension#
Algorithm: Newton's Method in One Dimension
We want to find a solution \(x^{\ast} \in I\) to the equation
where \(f: \mathcal{D} \subseteq \mathbb{R} \to \mathbb{R}\) is a real function and \(I \subseteq \mathcal{D}\) is some open interval.
- Pick an initial guess \(x_0 \in I\).
- Define the sequence \((x_n)_{n \in \mathbb{N}_0}\) as follows:
- Evaluate \((x_n)_{n \in \mathbb{N}_0}\) until \(|x_{n+1} - x_n| \lt \tau\), where \(\tau \gt 0\) is your acceptable error tolerance.
Provided that \(f\) satisfies some conditions and if the initial guess is chosen appropriately, the above algorithm converges quadratically to a root of \(f\).
Theorem: Local Quadratic Convergence
Let \(f: \mathcal{D} \subseteq \mathbb{R} \to \mathbb{R}\) be a real function and let \(I \subseteq \mathcal{D}\) be open interval.
If \(f\) satisfies the following conditions:
- It is twice continuously differentiable on \(I\);
- There exists some \(x^{\ast} \in I\) such that \(f(x^{\ast}) = 0\);
- \(f'(x^{\ast}) \ne 0\);
then there exists some open neighborhood \(N(x^{\ast})\) such that for any \(x_{0} \in N(x^{\ast})\) we have the following:
- \(f'(x_n) \ne 0\) for all \(n \ge 0\);
- The sequence \((x_n)_{n \in \mathbb{N}_0}\) with \(x_{n+1} = x_n - \frac{f(x_n)}{f'(x_n)}\) for all \(n \ge 0\) converges to \(x^{\ast}\);
- This convergence is quadratic:
Proof
TODO
Theorem: Global Monotone Convergence
Let \(f: \mathcal{D} \subseteq \mathbb{R} \to \mathbb{R}\) be a real function and let \([a,b] \subseteq \mathcal{D}\) be closed interval.
If \(f\) satisfies all of the following conditions:
- It is continuous on \([a,b]\) and twice continuously differentiable on \((a,b)\);
- It is monotone on \([a,b]\).
- There is some \(x^{\ast} \in (a,b)\) with \(f(x^{\ast}) = 0\);
- It is either strictly concave or strictly convex on \([a,b]\);
then for any \(x_0 \in [a,b]\) with \(f(x_0) \cdot f''(x_0) \gt 0\) we have the following:
- \(f'(x_n) \ne 0\) for all \(n \ge 0\);
- The sequence \((x_n)_{n \in \mathbb{N}_0}\) with \(x_{n+1} = x_n - \frac{f(x_n)}{f'(x_n)}\) for all \(n \ge 0\) is monotone and converges to \(x^{\ast}\).
Proof
TODO
In Multiple Dimensions#
Algorithm: Newton's Method
Let \(f: \mathcal{D} \subseteq \mathbb{R}^n \to \mathbb{R}^n\) be a real vector field which is totally differentiable on some open set \(U \subseteq \mathcal{D}\).
We want to find some \(\boldsymbol{p} \in U\) such that \(f(\boldsymbol{p}) = \boldsymbol{0}\).
- Pick an initial guess \(\boldsymbol{p}_0 \in U\), an iteration limit \(N_{\text{max}} \in \mathbb{N}\) and an acceptable error tolerance \(\tau \gt 0\).
- For \(i \in \{0, 1, \dotsc, N_{\text{max}}\}\):
- Attempt to solve \(J_f(\boldsymbol{p}_i) \Delta \boldsymbol{p}_i = -f(\boldsymbol{p}_i)\) for \(\Delta \boldsymbol{p}_i\). Return if an error occurs.
- If \(||\Delta \boldsymbol{p}_i|| \lt \tau\), then return \(\boldsymbol{p}_{i+1}\).
- Set \(\boldsymbol{p}_{i+1} = \boldsymbol{p}_i + \Delta \boldsymbol{p}_i\).
- If step 2 did not return, then return an error, since a desirable approximation of \(\boldsymbol{p}\) could not be found within the allotted number of iterations.
Provided that \(f\) satisfies some conditions and if the initial guess is chosen appropriately, the above algorithm converges quadratically to a root of \(f\).
Theorem: Local Quadratic Convergence (Multivariate Newton's Method)
Let \(f: \mathcal{D} \subseteq \mathbb{R}^n \to \mathbb{R}^n\) be a real vector field and let \(U \subseteq \mathcal{D}\) be an open set.
If \(f\) satisfies the following conditions:
- It is twice continuously differentiable on \(U\);
- There exists some \(x^{\ast} \in U\) such that \(f(\boldsymbol{x}^{\ast}) = \mathbf{0}\);
- The Jacobian \(J_f(\boldsymbol{x}^{\ast})\) is invertible;
then there exists some open neighborhood \(N(\boldsymbol{x}^{\ast})\) such that for any \(\boldsymbol{x}_{0} \in N(\boldsymbol{x}^{\ast})\) we have the following:
- \(J_f(\boldsymbol{x}_n)\) is invertible for all \(n \ge 0\);
- The sequence \((\boldsymbol{x}_n)_{n \in \mathbb{N}_0}\) defined as \(\boldsymbol{x}_{n+1} = \boldsymbol{x}_n - J^{-1}_f(\boldsymbol{x}_n)\cdot f(\boldsymbol{x}_n)\) for all \(n \ge 0\) converges to \(x^{\ast}\);
- This convergence is quadratic:
Proof
TODO