Markov chains

From Biology-Online Dictionary
Jump to: navigation, search

Markov chains

a stochastic process such that the conditional probability distribution for a state at any future instant, given the present state, is unaffected by any additional knowledge of the past history of the system.