|
179 | 179 | "cell_type": "markdown", |
180 | 180 | "metadata": {}, |
181 | 181 | "source": [ |
182 | | - "The attribute `msm.pi` tells us, for each discrete state, the absolute probability of observing said state in global equilibrium. Mathematically speaking, the stationary distribution $\\pi$ is the left eigenvector of the transition matrix $\\mathbf{T}$ to the eigenvalue $1$:\n", |
| 182 | + "The attribute `msm.pi` tells us, for each discrete state, the absolute probability of observing said state in global equilibrium.\n", |
| 183 | + "Mathematically speaking, the stationary distribution $\\pi$ is the left eigenvector of the transition matrix $\\mathbf{T}$ to the eigenvalue $1$:\n", |
183 | 184 | "\n", |
184 | 185 | "$$\\pi^\\top \\mathbf{T} = \\pi^\\top.$$\n", |
185 | 186 | "\n", |
186 | | - "We can use the stationary distribution to, e.g., visualize the weight of the dicrete states and, thus, to highlight which areas of our feature space are most probable. Here, we show all data points in a two dimensional scatter plot and color/weight them according to their discrete state membership:" |
| 187 | + "Please note that the $\\pi$ is fundamentaly different from a normalized histogram of states:\n", |
| 188 | + "for the histogram of states to accurately describe the stationary distribution, the data needs to be sampled from global equilibrium, i.e, the data points need to be statistically independent.\n", |
| 189 | + "The MSM approach, on the other hand, only requires local equilibrium, i.e., statistical independence of state transitions.\n", |
| 190 | + "Thus, the MSM approach requires a much weaker and, in practice, much easier to satisfy condition than simply counting state visits.\n", |
| 191 | + "\n", |
| 192 | + "We can use the stationary distribution to, e.g., visualize the weight of the dicrete states and, thus, to highlight which areas of our feature space are most probable.\n", |
| 193 | + "Here, we show all data points in a two dimensional scatter plot and color/weight them according to their discrete state membership:" |
187 | 194 | ] |
188 | 195 | }, |
189 | 196 | { |
|
236 | 243 | "cell_type": "markdown", |
237 | 244 | "metadata": {}, |
238 | 245 | "source": [ |
239 | | - "We will see further uses of the stationary distribution later. But for now, we continue the analysis of our model by visualizing its (right) eigenvectors. First, we notice that the first right eigenvector is a constant $1$." |
| 246 | + "We will see further uses of the stationary distribution later.\n", |
| 247 | + "But for now, we continue the analysis of our model by visualizing its (right) eigenvectors which encode the dynamical processes.\n", |
| 248 | + "First, we notice that the first right eigenvector is a constant $1$." |
240 | 249 | ] |
241 | 250 | }, |
242 | 251 | { |
|
281 | 290 | "cell_type": "markdown", |
282 | 291 | "metadata": {}, |
283 | 292 | "source": [ |
284 | | - "The right eigenvectors can be used to visualize the processes governed by the corresponding implied timescales. The first right eigenvector (always) is $(1,\\dots,1)^\\top$ for an MSM transition matrix and it corresponds to the stationary process (infinite implied timescale).\n", |
| 293 | + "The right eigenvectors can be used to visualize the processes governed by the corresponding implied timescales.\n", |
| 294 | + "The first right eigenvector (always) is $(1,\\dots,1)^\\top$ for an MSM transition matrix and it corresponds to the stationary process (infinite implied timescale).\n", |
285 | 295 | "\n", |
286 | | - "The second right eigenvector corresponds to the slowest process; its entries are negative for one group of discrete states and positive for the other group. This tells us that the slowest process happens between these two groups and that the process relaxes on the slowest ITS ($\\approx 8.5$ steps).\n", |
| 296 | + "The second right eigenvector corresponds to the slowest process;\n", |
| 297 | + "its entries are negative for one group of discrete states and positive for the other group.\n", |
| 298 | + "This tells us that the slowest process happens between these two groups and that the process relaxes on the slowest ITS ($\\approx 8.5$ steps).\n", |
287 | 299 | "\n", |
288 | | - "The third and fourth eigenvectors show a larger spread of values and no clear grouping. In combination with the ITS convergence plot, we can safely assume that these eigenvectors contain just noise and do not indicate any resolved processes.\n", |
| 300 | + "The third and fourth eigenvectors show a larger spread of values and no clear grouping.\n", |
| 301 | + "In combination with the ITS convergence plot, we can safely assume that these eigenvectors contain just noise and do not indicate any resolved processes.\n", |
289 | 302 | "\n", |
290 | 303 | "We then continue to validate our MSM with a CK test for $2$ metastable states which are already indicated by the second right eigenvector." |
291 | 304 | ] |
|
423 | 436 | "source": [ |
424 | 437 | "Again, we have the $(1,\\dots,1)^\\top$ first right eigenvector of the stationary process.\n", |
425 | 438 | "\n", |
426 | | - "The second to fourth right eigenvectors illustrate the three slowest processes.\n", |
| 439 | + "The second to fourth right eigenvectors illustrate the three slowest processes which are (in that order):\n", |
| 440 | + "\n", |
| 441 | + "- rotation of the $\\Phi$ dihedral,\n", |
| 442 | + "- rotation of the $\\Psi$ dihedral when $\\Phi\\approx-2$ rad, and\n", |
| 443 | + "- rotation of the $\\Psi$ dihedral when $\\Phi\\approx1$ rad.\n", |
427 | 444 | "\n", |
428 | 445 | "Eigenvectors five, six, and seven indicate further processes which, however, relax faster than the lag time and cannot be resolved clearly.\n", |
429 | 446 | "\n", |
|
0 commit comments