You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
On $\mathbf{g}$ we then apply fixed-point iteration.
235
+
On ${g}$ we then apply fixed-point iteration.
236
236
237
237
We saw one example how to achieve this rewriting in the discussion
238
238
of the "Intersecting circle and parabola" example in the [section above](#Root-finding-and-fixed-point-problems), where we first defined a root-finding problem and then two equivalent fixed-point problems for the same task.
@@ -355,7 +355,7 @@ See also the discussion in [Revision and preliminaries](https://teaching.matmat.
355
355
Sometimes being fully precise in the big O notation will be too distracting. In this case we will use a generic "$O(\text{small})$" to remind ourselves that there are additional terms and we will specify in the surrounding text what this term stands for.
356
356
"""
357
357
358
-
# ╔═╡ 01db98ec-daf2-4779-9f31-c3271039f44c
358
+
# ╔═╡ 7ca9192e-e584-480f-8d20-ac8fe3e3d46d
359
359
md"""
360
360
Using (3), the fact that $g(x_\ast) = x_\ast$,
361
361
and the key fixed-point iterations equation, $x^{(k+1)} = g(x^{(k)})$,
In other words as $k \to \infty$, i.e. the iteration progresses,
393
-
$|e^{(k+1)}|$ approaches zero
394
-
if $|g'(x_\ast)| < 1$.
397
+
where the $O(\text{small})$ is at least quadratic in all errors $|e^{(k)}|$ to $|e^{(0)}|$.
398
+
"""
399
+
400
+
# ╔═╡ 15c5b579-25f2-4d37-960b-031e80a7a1aa
401
+
md"""
402
+
From this we conclude:
403
+
- If $|g'(x_\ast)| > 1$, then $|g'(x_\ast)|^k$ grows to infinity as the number of iteration $k$ grows, therefore the error $|e^{(k+1)}|$ has to grow: **the fixed-point iterations diverge.**
404
+
- Furthermore if $|g'(x_\ast)| < 1$ and if the terms hidden in $O(\text{small})$ can be neglected, then $|e^{(k+1)}|$ is getting smaller and smaller as the number of iterations $k$ increases: **the fixed-point iterations converge**.
405
+
- Now we ask under which conditions the terms in $O(\text{small})$ can be neglected. This is indeed the case if $|e^{(0)}|$ is sufficiently small, i.e. **if our starting point $x^{(0)}$ is sufficiently close to the fixed point $x_\ast$**.
406
+
- To see this, assume $|g'(x_\ast)| < 1$. As a result $|e^{(1)}| < |g'(x_\ast)| \ |e^{(0)}| + O(|e^{(0)}|^2)$ by ($\ast$) for $k=1$. Therefore if $|e^{(0)}|$ is sufficiently small, then $|e^{(0)}|^2 = |x_\ast - x^{(0)}|^2$ is small compared to $|g'(x_\ast)| \ |e^{(0)}|$, meaning that the terms hidden in $O(|e^{(0)}|^2)$ can be neglected. As a result $|e^{(1)}|^2$ is even smaller than $|e^{(0)}|^2$, such that also the terms in $O(|e^{(1)}|^2)$ can be neglected and so forth. We conclude that all terms at least quadratic in the error term $|e^{(k)}|$ to $|e^{(0)}|$ can be neglected, i.e. that the entire set of terms $O(\text{small})$ is neglibile.
407
+
- Note that for $|g'(x_\ast)| = 1$ our theory does not allow us to conclude neither convergence, nor divergence.
408
+
409
+
We summarise in a Theorem:
395
410
"""
396
411
397
412
# ╔═╡ 9176b666-41f7-436e-b5ad-61b196a8b35b
398
413
md"""
399
-
!!! note "Theorem 1 (scalar version)"
414
+
!!! note "Theorem 0 (scalar version of Theorem 1)"
400
415
Let $g : \mathbb{R} \to \mathbb{R}$
401
416
be a once differentiable function
402
417
and $x_\ast \in \mathbb{R}$ be a fixed point of $g$.
@@ -414,7 +429,7 @@ We will generalise this theorem to the vector case in the following secition.
414
429
md"""
415
430
#### Higher dimensions
416
431
417
-
We now consider the generalisation of the above argument to the vector setting,
432
+
We now consider the generalisation of the above argument to the multi-dimensional setting,
418
433
i.e. finding a fixed-point $\mathbf{x}_\ast = {g}(\mathbf{x}_\ast) \in \mathbb{R}^n$ of a function ${g} : \mathbb{R}\to\mathbb{R}$.
419
434
To make a similar argument to the scalar case, we need to consider again the Talyor expansion of ${g}(\mathbf{x}^{(k)}) = {g}(\mathbf{x}_\ast + \mathbf{e}^{(k)})$ around $\mathbf{x}_\ast$, where as before $\mathbf{e}^{(k)} = \mathbf{x}^{(k)} - \mathbf{x}_\ast$.
420
435
@@ -450,9 +465,8 @@ is the collection of all partial derivatives of ${g}$ *evaluated at $\mathbf{x}$
450
465
\end{array}\right).
451
466
```
452
467
See also the discussion on multi-dimensional Talyor approximations in [Revision and preliminaries](https://teaching.matmat.org/numerical-analysis/03_Preliminaries.html).
453
-
454
468
Note that the Jacobian (just like any derivative) is a function of an independent
455
-
variable (here $\textbf{x}_\ast$).
469
+
variable (here $\textbf{x}$).
456
470
457
471
Since the Jacobian very much plays the role of a generalised derivative
458
472
of a multidimensional function ${g}$, we will sometimes also
@@ -547,7 +561,6 @@ The definition of the matrix norm implies in particular that
547
561
# ╔═╡ 9c9719e3-ec6c-4bdc-b05b-ab4bd4119cb9
548
562
md"""
549
563
We now take vector norms on either side of (5) and make use of this last inequality to obtain to first order
where $O(\text{small})$ is a small term that we do not make more precise for simplicity.
563
576
564
-
We are again faced with the conclusion that as $k \to \infty$, i.e. the iteration progresses, that the error norm $\|\mathbf{e}^{(k+1)}\|$ approaches zero
565
-
if $\left\|\mathbf{J}_{g}(\mathbf{x}_\ast) \right\| < 1$.
577
+
Similar to the one-dimensional case we conclude:
578
+
- If $\left\|\mathbf{J}_{g}(\mathbf{x}_\ast) \right\| < 1$ and the initial guess $\mathbf{x}^{(0)}$ is sufficiently close to the final fixed point $\mathbf{x}_\ast$, then as $k \to \infty$, i.e. as the iteration progresses, the error norm $\|\mathbf{e}^{(k+1)}\|$ approaches zero: **the fixed-point iterations converge**.
579
+
580
+
However, in contrast to the one-dimensional setting **we cannot conclude divergence** if $\left\|\mathbf{J}_{g}(\mathbf{x}_\ast) \right\| > 1$. This is because we only obtain an *inequality* relating $\|\mathbf{e}^{(k+1)}\|$ to $\|\mathbf{e}^{(0)}\|$, whereas in the one-dimensional case this equation was an equality.
566
581
567
582
The following theorem summarises our argument
568
583
"""
@@ -722,7 +737,7 @@ In a given step $\mathbf{x}^{(k)}$ we have in general not yet achieved our goal,
722
737
i.e. ${g}(\mathbf{x}^{(k)}) \neq \mathbf{x}^{(k)}$.
723
738
An idea is thus to consider exactly the descrepancy
Note, that this is just a **conceptional expression** as determining
829
841
$\xi^{(k)}$ is in general *as hard* as finding $x_\ast$.
830
842
But it will be useful in some theoretical arguments.
831
-
- For converging fiterations $x^{(k)} \to x_\ast$ as $k \to \infty$.
843
+
- For converging iterations $x^{(k)} \to x_\ast$ as $k \to \infty$.
832
844
Therefore the interval $[x_\ast, x^{(k)}]$ gets smaller and smaller,
833
845
such that necessarily $\xi^{(k)} \to x_\ast$
834
846
and $g'(\xi^{(k)}) \to g'(x_\ast)$ as $k \to \infty$.
@@ -1024,7 +1036,7 @@ md"""
1024
1036
#### Visual inspection: Residual ratio
1025
1037
1026
1038
One caveat with this analysis is that we cheated a little by assuming that we already *know* the solution. An alternative approach is to **build upon our
1027
-
residual-error relationship**, i.e. for the scalar case (5)
1039
+
residual-error relationship**, i.e. for the scalar case (6)
0 commit comments