Skip to content

Commit e2bc5c1

Browse files
committed
More md -> mdx conversions.
1 parent e7ae14d commit e2bc5c1

9 files changed

+400
-479
lines changed

src/content/post/brouwers-fixed-point.md

Lines changed: 0 additions & 54 deletions
This file was deleted.
Lines changed: 55 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,55 @@
1+
---
2+
title: "Brouwer's Fixed Point Theorem: A Proof with Reduced Homology"
3+
description: "A proof of special case (ball) of Brouwer's Fixed Point Theorem with Reduced Homology."
4+
publishDate: 2018-07-18 10:00
5+
tags: [math]
6+
---
7+
import BlogImage from '@/components/BlogImage.astro'
8+
9+
This post is about the proof I found very interesting during the Topology course I took this semester. It highlights the application of Reduced Homology, which is a modification of Homology theory in Algebraic Topology. We will use two results from Reduced Homology as black-boxes for the proof. Everywhere, we will assume $ \mathbb{Q} $ is used as the coefficient of the Homology space.
10+
11+
**Lemma 1 (Reduced Homology of spheres)**
12+
Given a $ d $-sphere $ \mathbb{S}^d $, then its reduced $ p $-th Homology space is:
13+
14+
$$
15+
\tilde{H}_p(\mathbb{S}^d) = \begin{cases} \mathbb{Q}, & \text{if } p = d \\ 0, & \text{otherwise} \enspace . \end{cases}
16+
$$
17+
18+
$$
19+
\qed
20+
$$
21+
22+
**Lemma 2 (Reduced Homology of balls)**
23+
Given a $ d $-ball $ \mathbb{B}^d $, then its reduced $ p $-th Homology space is trivial, i.e. $\tilde{H}_p(\mathbb{B}^d) = 0 $, for any $ d $ and $ p $.
24+
25+
$$
26+
\qed
27+
$$
28+
29+
Equipped with these lemmas, we are ready to prove the special case of Brouwer's Fixed Point Theorem, where we consider map from a ball to itself.
30+
31+
**Brouwer's Fixed Point Theorem**
32+
Given $ f: \mathbb{B}^{d+1} \to \mathbb{B}^{d+1} $ continuous, then there exists $ x
33+
\in \mathbb{B}^{d+1} $ such that $ f(x) = x $.
34+
35+
_Proof._    For contradiction, assume $ \forall x \in \mathbb{B}^{d+1}: f(x) \neq x $. We construct a map $ r: \mathbb{B}^{d+1} \to \mathbb{S}^d $, casting ray from the ball to its shell by extending the line segment between $ x $ and $ f(x) $.
36+
37+
<BlogImage imagePath='/img/brouwers-fixed-point/map_r.svg' altText="Casting a ray to the ball's shell." />
38+
39+
Observe that $ r(x) $ is continuous because $ f(x) $ is. Also, $ x \in \mathbb{S}^d \implies r(x) = x $. Therefore we have the following commutative diagram.
40+
41+
<BlogImage imagePath='/img/brouwers-fixed-point/comm_diag.svg' altText="Commutative diagram." />
42+
43+
Above, $ i $ is inclusion map, and $ id $ is identity map. We then look of the Reduced Homology of the above, and this gives us the following commutative diagram.
44+
45+
<BlogImage imagePath='/img/brouwers-fixed-point/comm_diag_hom.svg' altText="Reduced homology." />
46+
47+
As the diagram commute, then $ \tilde{H}_d(\mathbb{S}^d) \xrightarrow{i^*} \tilde{H}_d(\mathbb{B}^{d+1}) \xrightarrow{r^*} \tilde{H}_d(\mathbb{S}^d) $ should be identity map on $ \tilde{H}_d(\mathbb{S}^d) $. By Lemma 2, $ \tilde{H}_d(\mathbb{B}^{d+1}) = 0 $. This implies $ \tilde{H}_d(\mathbb{S}^d) = 0 $. But this is a contradiction, as By Lemma 1, $ \tilde{H}_d(\mathbb{S}^d) = \mathbb{Q} $. Therefore there must be a fixed point.
48+
49+
$$
50+
\qed
51+
$$
52+
53+
## References
54+
55+
1. Hatcher, Allen. "Algebraic topology." (2001).
Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ publishDate: 2018-03-11 07:00
55
tags: [machine learning]
66
---
77

8-
Suppose we have a model parameterized by parameter vector \\( \theta \\) that models a distribution \\( p(x \vert \theta) \\). In frequentist statistics, the way we learn \\( \theta \\) is to maximize the likelihood \\( p(x \vert \theta) \\) wrt. parameter \\( \theta \\). To assess the goodness of our estimate of \\( \theta \\) we define a score function:
8+
Suppose we have a model parameterized by parameter vector $ \theta $ that models a distribution $ p(x \vert \theta) $. In frequentist statistics, the way we learn $ \theta $ is to maximize the likelihood $ p(x \vert \theta) $ wrt. parameter $ \theta $. To assess the goodness of our estimate of $ \theta $ we define a score function:
99

1010
$$
1111
s(\theta) = \nabla_{\theta} \log p(x \vert \theta) \, ,
@@ -16,7 +16,7 @@ that is, score function is the gradient of log likelihood function. The result a
1616
**Claim:**
1717
The expected value of score wrt. our model is zero.
1818

19-
_Proof._ &nbsp;&nbsp; Below, the gradient is wrt. \\( \theta \\).
19+
_Proof._ &nbsp;&nbsp; Below, the gradient is wrt. $ \theta $.
2020

2121
$$
2222
\begin{align}
@@ -40,13 +40,13 @@ $$
4040
\mathop{\mathbb{E}}_{p(x \vert \theta)} \left[ (s(\theta) - 0) \, (s(\theta) - 0)^{\text{T}} \right] \, .
4141
$$
4242

43-
We can then see it as an information. The covariance of score function above is the definition of Fisher Information. As we assume \\( \theta \\) is a vector, the Fisher Information is in a matrix form, called Fisher Information Matrix:
43+
We can then see it as an information. The covariance of score function above is the definition of Fisher Information. As we assume $ \theta $ is a vector, the Fisher Information is in a matrix form, called Fisher Information Matrix:
4444

4545
$$
4646
\text{F} = \mathop{\mathbb{E}}_{p(x \vert \theta)} \left[ \nabla \log p(x \vert \theta) \, \nabla \log p(x \vert \theta)^{\text{T}} \right] \, .
4747
$$
4848

49-
However, usually our likelihood function is complicated and computing the expectation is intractable. We can approximate the expectation in \\( \text{F} \\) using empirical distribution \\( \hat{q}(x) \\), which is given by our training data \\( X = \\{ x_1, x_2, \cdots, x_N \\} \\). In this form, \\( \text{F} \\) is called Empirical Fisher:
49+
However, usually our likelihood function is complicated and computing the expectation is intractable. We can approximate the expectation in $ \text{F} $ using empirical distribution $ \hat{q}(x) $, which is given by our training data $ X = \{ x_1, x_2, \cdots, x_N \} $. In this form, $ \text{F} $ is called Empirical Fisher:
5050

5151
$$
5252
\begin{align}
@@ -56,10 +56,10 @@ $$
5656

5757
## Fisher and Hessian
5858

59-
One property of \\( \text{F} \\) that is not obvious is that it has the interpretation of being the negative expected Hessian of our model's log likelihood.
59+
One property of $ \text{F} $ that is not obvious is that it has the interpretation of being the negative expected Hessian of our model's log likelihood.
6060

6161
**Claim:**
62-
The negative expected Hessian of log likelihood is equal to the Fisher Information Matrix \\( \text{F} \\).
62+
The negative expected Hessian of log likelihood is equal to the Fisher Information Matrix $ \text{F} $.
6363

6464
_Proof._ &nbsp;&nbsp; The Hessian of the log likelihood is given by the Jacobian of its gradient:
6565

@@ -85,19 +85,19 @@ $$
8585
\end{align}
8686
$$
8787

88-
Thus we have \\( \text{F} = -\mathop{\mathbb{E}}\_{p(x \vert \theta)} \left[ \text{H}\_{\log p(x \vert \theta)} \right] \\).
88+
Thus we have $ \text{F} = -\mathop{\mathbb{E}}\_{p(x \vert \theta)} \left[ \text{H}\_{\log p(x \vert \theta)} \right] $.
8989

9090
$$
9191
\qed
9292
$$
9393

94-
Indeed knowing this result, we can see the role of \\( \text{F} \\) as a measure of curvature of the log likelihood function.
94+
Indeed knowing this result, we can see the role of $ \text{F} $ as a measure of curvature of the log likelihood function.
9595

9696
## Conclusion
9797

98-
Fisher Information Matrix is defined as the covariance of score function. It is a curvature matrix and has interpretation as the negative expected Hessian of log likelihood function. Thus the immediate application of \\( \text{F} \\) is as drop-in replacement of \\( \text{H} \\) in second order optimization methods.
98+
Fisher Information Matrix is defined as the covariance of score function. It is a curvature matrix and has interpretation as the negative expected Hessian of log likelihood function. Thus the immediate application of $ \text{F} $ is as drop-in replacement of $ \text{H} $ in second order optimization methods.
9999

100-
One of the most exciting results of \\( \text{F} \\) is that it has connection to KL-divergence. This gives rise to natural gradient method, which we shall discuss further in the next article.
100+
One of the most exciting results of $ \text{F} $ is that it has connection to KL-divergence. This gives rise to natural gradient method, which we shall discuss further in the next article.
101101

102102
## References
103103

src/content/post/mayer-vietoris-sphere.md

Lines changed: 0 additions & 55 deletions
This file was deleted.
Lines changed: 60 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,60 @@
1+
---
2+
title: 'Reduced Betti number of sphere: Mayer-Vietoris Theorem'
3+
description: 'A proof of reduced homology of sphere with Mayer-Vietoris sequence.'
4+
publishDate: 2018-07-23 10:00
5+
tags: [math]
6+
---
7+
8+
import BlogImage from '@/components/BlogImage.astro'
9+
10+
In the previous post about Brouwer's Fixed Point Theorem, we used two black boxes. In this post we will prove the slight variation of those black boxes. We will start with the simplest lemma first: the reduced homology of balls.
11+
12+
**Lemma 2 (Reduced homology of balls)**
13+
Given a $ d $-ball $ \mathbb{B}^d $, then its reduced $ p $-th homology space is trivial, i.e. $\tilde{H}_p(\mathbb{B}^d) = 0 $, for any $ d $ and $ p $.
14+
15+
_Proof._ &nbsp;&nbsp; Observe that $ \mathbb{B}^d $ is contractible, i.e. homotopy equivalent to a point. Assuming we use coefficient $ \mathbb{Q} $, we know the zero-th homology space of point is $ H_0(\, \cdot \,, \mathbb{Q}) = \mathbb{Q} $, and trivial otherwise, i.e. $ H_p (\, \cdot \,, \mathbb{Q}) = 0 \enspace \forall p \geq 1 $.
16+
17+
In the reduced homology, therefore $ \tilde{H}_0(\, \cdot \,, \mathbb{Q}) = 0 $. Thus the reduced homology of balls is trivial for all $ d, p $.
18+
19+
$$
20+
\qed
21+
$$
22+
23+
**Corollary 1 (Reduced Betti numbers of balls)**
24+
The $ p $-th reduced Betti numbers of $ \mathbb{B}^d $ is zero for all $d, p$.
25+
26+
$$
27+
\qed
28+
$$
29+
30+
Now, we are ready to prove the main theme of this post.
31+
32+
**Lemma 1 (Reduced Betti numbers of spheres)**
33+
Given a $ d $-sphere $ \mathbb{S}^d $, then its $ p $-th reduced Betti number is:
34+
35+
$$
36+
\tilde{\beta}_p(\mathbb{S}^d) = \begin{cases} 1, & \text{if } p = d \\ 0, & \text{otherwise} \enspace . \end{cases}
37+
$$
38+
39+
_Proof._ &nbsp;&nbsp; We use "divide-and-conquer" approach to apply Mayer-Vietoris Theorem. We cut the sphere along the equator and note that the upper and lower portion of the sphere is just a disk, and the intersection between those two parts is a circle (sphere one dimension down), as shown in the figure below.
40+
41+
<BlogImage
42+
imagePath='/img/mayer-vietoris-sphere/sphere.svg'
43+
altText='Mayer-Vietoris on a sphere.'
44+
/>
45+
46+
By Mayer-Vietoris Theorem, we have a long exact sequence in the form of:
47+
48+
$$
49+
\dots \longrightarrow \tilde{H}_p(\mathbb{S}^{d-1}) \longrightarrow \tilde{H}_p(\mathbb{B}^d) \oplus \tilde{H}_p(\mathbb{B}^d) \longrightarrow \tilde{H}_p(\mathbb{S}^d) \longrightarrow \tilde{H}_{p-1}(\mathbb{S}^{d-1}) \longrightarrow \dots \enspace .
50+
$$
51+
52+
By Corollary 1, $ \tilde{H}_p(\mathbb{B}^d) \oplus \tilde{H}_p(\mathbb{B}^d) = \tilde{H}_{p-1}(\mathbb{B}^d) \oplus \tilde{H}_{p-1}(\mathbb{B}^d) = 0 $. As the sequence is exact, therefore $ \tilde{H}_p(\mathbb{S}^d) \longrightarrow \tilde{H}_{p-1}(\mathbb{S}^{d-1}) $ is a bijection, and thus an isomorphism. Then by induction with base case of $ \mathbb{S}^0 $, we conclude that the claim holds.
53+
54+
$$
55+
\qed
56+
$$
57+
58+
## References
59+
60+
1. Hatcher, Allen. "Algebraic topology." (2001).

0 commit comments

Comments
 (0)