Description Usage Format Examples
A transition probaility matrix for understanding Markov chains.
1 |
A matrix of transition probability matrix
A
transitions probabilities from State A
B
transitions probabilities from State B
C
transitions probabilities from State C
D
transitions probabilities from State D
E
transitions probabilities from State E
F
transitions probabilities from State F
1 |
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.