Skip to content

Commit 3fb4617

Browse files
committed
Fix footnotes in probabilistic PCA tutorial
1 parent 81a8412 commit 3fb4617

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

tutorials/11-probabilistic-pca/index.qmd

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -278,12 +278,12 @@ Another way to put it: 2 dimensions is enough to capture the main structure of t
278278
A direct question arises from above practice is: how many principal components do we want to keep, in order to sufficiently represent the latent structure in the data?
279279
This is a very central question for all latent factor models, i.e. how many dimensions are needed to represent that data in the latent space.
280280
In the case of PCA, there exist a lot of heuristics to make that choice.
281-
For example, We can tune the number of principal components using empirical methods such as cross-validation based some criteria such as MSE between the posterior predicted (e.g. mean predictions) data matrix and the original data matrix or the percentage of variation explained [3].
281+
For example, We can tune the number of principal components using empirical methods such as cross-validation based some criteria such as MSE between the posterior predicted (e.g. mean predictions) data matrix and the original data matrix or the percentage of variation explained [^3].
282282

283283
For p-PCA, this can be done in an elegant and principled way, using a technique called *Automatic Relevance Determination* (ARD).
284-
ARD can help pick the correct number of principal directions by regularizing the solution space using a parameterized, data-dependent prior distribution that effectively prunes away redundant or superfluous features [4].
284+
ARD can help pick the correct number of principal directions by regularizing the solution space using a parameterized, data-dependent prior distribution that effectively prunes away redundant or superfluous features [^4].
285285
Essentially, we are using a specific prior over the factor loadings $\mathbf{W}$ that allows us to prune away dimensions in the latent space. The prior is determined by a precision hyperparameter $\alpha$. Here, smaller values of $\alpha$ correspond to more important components.
286-
You can find more details about this in e.g. [5].
286+
You can find more details about this in, for example, Bishop (2006) [^5].
287287

288288
```{julia}
289289
@model function pPCA_ARD(X)

0 commit comments

Comments
 (0)