# Markov chain

From Conservapedia

A **Markov chain**, sometimes called a Markov process, is a sequence in which any state in the sequence depends only on the previous state, and is independent of all other states. An example of a Markov process is the random walk problem.