|
72 | 72 | "cell_type": "markdown", |
73 | 73 | "metadata": {}, |
74 | 74 | "source": [ |
75 | | - "Given the low dimensionality of this data set, we can directly attempt to discretize, e.g. with $k$-means with $100$ centers and a stride of $5$ to reduce the computational effort..." |
| 75 | + "Given the low dimensionality of this data set, we can directly attempt to discretize, e.g. with $k$-means with $100$ centers and a stride of $5$ to reduce the computational effort. In real world examples we also might encounter low dimensional feature spaces, which do not require further dimension reduction techniques to be clustered efficiently." |
76 | 76 | ] |
77 | 77 | }, |
78 | 78 | { |
|
129 | 129 | "cell_type": "markdown", |
130 | 130 | "metadata": {}, |
131 | 131 | "source": [ |
132 | | - "Have you noticed how the $k$-means centers follow the density of the data points while the regspace centers are spread uniformly over the whole area?\n", |
| 132 | + "Have you noticed how the $k$-means centers follow the density of the data points while the regspace centers are spread uniformly over the whole area? \n", |
| 133 | + "\n", |
| 134 | + "If your are only interested in well sampled states, you should use a density based method to discretize. If exploring new states is one of your objectives, it might be of advantage to place states also in rarely observed regions. The latter is especially useful in adaptive sampling approaches, because in the initial phase you want to explore the phase space as much as possible. The downside of placing states in areas of low density is that we will have poor statistics on these states. \n", |
| 135 | + "\n", |
| 136 | + "Another advantage of regular space clustering is that it is very fast in comparison to $k$-means: regspace clustering runs in linear time while $k$-means is superpolynomial in time.\n", |
| 137 | + "\n", |
| 138 | + "For very large datasets we also offer a mini batch version of $k$-means, which has the same semantics as the original method, but trains the centers on subsets of your data. This tutorial does not cover this case, but you should keep in mind, that $k$-means requires your low dimensional space to fit into your main memory.\n", |
133 | 139 | "\n", |
134 | 140 | "The main result of a discretization for Markov modeling, however, is not the set of centers but the time series of discrete states. These are accessible via the `dtrajs` attribute of any clustering object:" |
135 | 141 | ] |
|
372 | 378 | "source": [ |
373 | 379 | "Again, notice the difference between $k$-means and regspace clustering.\n", |
374 | 380 | "\n", |
375 | | - "Now, we use a different featurization for the same data set and revisit how to use PCA and TICA.\n", |
| 381 | + "Now, we use a different featurization for the same data set and revisit how to use PCA and TICA. In practice you almost never would like to use PCA as dimension reduction method in MSM building, as it does not preserve kinetic variance. We are showing it here in these exercises to make this point clear.\n", |
376 | 382 | "\n", |
377 | 383 | "#### Exercise 1: data loading \n", |
378 | 384 | "\n", |
|
1180 | 1186 | "name": "python", |
1181 | 1187 | "nbconvert_exporter": "python", |
1182 | 1188 | "pygments_lexer": "ipython3", |
1183 | | - "version": "3.6.5" |
| 1189 | + "version": "3.6.6" |
1184 | 1190 | }, |
1185 | 1191 | "toc": { |
1186 | 1192 | "base_numbering": 1, |
|
0 commit comments