In this section, we will only consider single variable functions. The proof for multivariable functions is similar except one would have to use inverse matrices of the Jacobian instead of division by f ' making proof much more difficult to read.
f(r) = f(xk + ek) = f(xk) + f'(xk)ek + f"(ξk)ek2 / 2where ξk is between xk and r. But f(r) = 0 so
0 = f(xk) + f'(xk)ek + f"(ξk)ek2 / 2Divide by f'(xk) to obtain
0 = f(xk) / f'(xk) + ek + f"(ξk)ek2 / [2 f'(xk)]Now replace ek by its value obtained earlier.
0 = f(xk) / f'(xk) + ek+1 - f(xk) / f'(xk) + f"(ξk)ek2 / [2 f'(xk)] 0 = ek+1 + f"(ξk)ek2 / [2 f'(xk)] ek+1 = {-f"(ξk) / [2 f'(xk)]} ek2If we maximize -f"(ξk) / [2 f'(xk) in an appropriate interval near r and apply absolute values, we see that the convergence is quadratic if x0 is close enough to r.
(Again in this section, we will only consider single variable functions. The proof for multivariable functions is similar except one would have to use inverse matrices of the Jacobian instead of division by f ' making proof much more difficult to read.)
Mathematical analysis of the approximation method shows almost quadratic convergence providing f '(x) ≠ 0 at the solution. It also shows that there are serious problems if f '(x) = 0 at the solution. This analysis follows the proof that Newton's method converges quadratically. Changes in the above proof are shown in italics.d(xk, Δ) = f'(xk + Δ) - f'(xk) Δand using the iteration
f(r) = f(xk - ek) = f(xk) - f'(xk)ek + f"(ξk)ek2/2where ξk is between xk and r. But f(r) = 0 so
0 = f(xk) - f'(xk)ek + f"(ξk)ek2/2Divide by f'(xk) to obtain
0 = f(xk) / f'(xk) + ek + f"(ξk)ek2 / [2 f'(xk)]Now replace ek by its value obtained earlier.
0 = f(xk) / f'(xk) + ek+1 - f(xk) / d(xk, Δ) + f"(ξk)ek2 / [2 f'(xk)] 0 = ek+1 + f(xk) / f'(xk) - f(xk) / d(xk, Δ) + f"(ξk)ek2 / [2 f'(xk)] ek+1 = -f(xk) / f'(xk) + f(xk) / d(xk, Δ) + {-f"(ξk) / [2 f'(xk)]} ek2 = -f(xk) * [1 / f'(xk) - 1 / d(xk, Δ)] + {-f"(ξk) / [2 f'(xk)]} ek2 = Q(xk, δk) + Ck ek2where
Q(xk, Δ) = -f(xk) * [1 / f'(xk) - 1 / d(xk, Δ)]and
Ck = -f"(ξk) 2f'(xk)Consider Q(xk, Δ). If Q(xk, Δ) is zero, then the method would converge quadratically. But in general, it will not be. Lets examine [1 / f'(xk) - 1 / d(xk, Δ)] more closely. Because
lim Q(xk, Δ) Δ → 0 = lim f(x+Δ) - f(x) = f'(x) Δ → 0 ΔWe see
lim [1 / f'(xk) - 1 / d(xk, Δ)] = 0 Δ → 0Thus we can make Q(xk) as close to 0 as desired by making Δ small enough. Unfortunately, in practice, we will lose numerical accuracy if it is too small. This would be a significant problem if we were using single precision arithmetic. But in modern computers and many languages (including Javascript), double precision is used without much inefficiency. (This technique would not be recommended if calculations are carried out in single precision.) Experimentally it has been found that using Δ from 10-4 to 10-6 often approximates the derivative very well and the resulting convergence appears to be quadratic. (Sometimes, negative values of Δ can cause faster convergence.) Even smaller values have been used successfully.
The expression for Q(xk, Δ) also shows that there can be serious problems with convergence if f '(r) = 0 because both f '(xk) and d(xk, Δ) approach 0 as xk approaches the solution. While iteration patterns when f '(r) ≠ 0 normally closely resemble those of regular Newton's method, they often differ when f '(r) = 0 as the iterations start getting closer r.
X = ⌈x⌉ F(x, y) = ⌈f0(x, y)⌉ E = ⌈x - r0⌉ ⌊y⌋ ⌊f1(x, y)⌋ ⌊y - r1⌋ J(x, y) = ⌈δf0(x, y)/δx δf0(x, y)/δy⌉ ⌊δf1(x, y)/δx δf1(x, y)/δy⌋Then Newton's method can be carried out with
Xk+1 = Xk - J(xk, yk)-1F(xk, yk)Actually it is not necessary to calculate J(xk, yk)-1. Instead one can solve
δf0(xk, yk)/δx = f0(xk + Δ, yk) - f0(xk, yk) Δ δf0(xk, yk)/δy = f0(xk, yk + Δ) - f0(xk, yk) ΔLikewise for f1 when approximating Newton's method. The same concept is used for systems with more variables. Experience has shown that Δ = 0.0001 often produces quadratic like convergence.
©James Brink, 11/12/2021