You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Here $\Sigma G' (G \Sigma G' + R)^{-1}$ is the matrix of population regression coefficients of the hidden object $x - \hat{x}$ on the surprise $y - G \hat{x}$.
284
+
Here $\Sigma G' (G \Sigma G' + R)^{-1}$ is the matrix of population regression coefficients of the hidden object $x - \hat{x}$ on the surprise $y - G \hat{x}$.
285
+
286
+
We can verify it by computing
287
+
```{math}
288
+
\mathrm{Cov}(x - \hat{x}, y - G \hat{x})\mathrm{Var}(y - G \hat{x})^{-1}
289
+
= \mathrm{Cov}(x - \hat{x}, G x + v - G \hat{x})\mathrm{Var}(G x + v - G \hat{x})^{-1}
290
+
= \Sigma G'(G \Sigma G' + R)^{-1}
291
+
```
288
292
289
293
This new density $p(x \,|\, y) = N(\hat{x}^F, \Sigma^F)$ is shown in the next figure via contour lines and the color map.
290
294
@@ -343,7 +347,7 @@ We have obtained probabilities for the current location of the state (missile) g
343
347
This is called "filtering" rather than forecasting because we are filtering
344
348
out noise rather than looking into the future.
345
349
346
-
* $p(x \,|\, y) = N(\hat x^F, \Sigma^F)$ is called the *filtering distribution*
350
+
* $p(x \,|\, y) = N(\hat x^F, \Sigma^F)$ is called the **filtering distribution**
347
351
348
352
But now let's suppose that we are given another task: to predict the location of the missile after one unit of time (whatever that may be) has elapsed.
349
353
@@ -382,7 +386,7 @@ $$
382
386
$$
383
387
384
388
The matrix $A \Sigma G' (G \Sigma G' + R)^{-1}$ is often written as
385
-
$K_{\Sigma}$ and called the *Kalman gain*.
389
+
$K_{\Sigma}$ and called the **Kalman gain**.
386
390
387
391
* The subscript $\Sigma$ has been added to remind us that $K_{\Sigma}$ depends on $\Sigma$, but not $y$ or $\hat x$.
388
392
@@ -399,7 +403,7 @@ Our updated prediction is the density $N(\hat x_{new}, \Sigma_{new})$ where
399
403
\end{aligned}
400
404
```
401
405
402
-
* The density $p_{new}(x) = N(\hat x_{new}, \Sigma_{new})$ is called the *predictive distribution*
406
+
* The density $p_{new}(x) = N(\hat x_{new}, \Sigma_{new})$ is called the **predictive distribution**
403
407
404
408
The predictive distribution is the new density shown in the following figure, where
405
409
the update has used parameters.
@@ -743,7 +747,7 @@ plt.show()
743
747
As discussed {ref}`above <kalman_convergence>`, if the shock sequence $\{w_t\}$ is not degenerate, then it is not in general possible to predict $x_t$ without error at time $t-1$ (and this would be the case even if we could observe $x_{t-1}$).
744
748
745
749
Let's now compare the prediction $\hat x_t$ made by the Kalman filter
746
-
against a competitor who **is** allowed to observe $x_{t-1}$.
750
+
against a competitor who *is* allowed to observe $x_{t-1}$.
747
751
748
752
This competitor will use the conditional expectation $\mathbb E[ x_t
749
753
\,|\, x_{t-1}]$, which in this case is $A x_{t-1}$.
0 commit comments