Definitions for "Markov process"
a random process in which the probabilities of states in a series depend only on the properties of the immediately preceding state or the next preceeding state, independent of the path by which the preceding state was reached. It is distinguished from a Markov chain in that the states of a Markov process may be continuous as well as discrete.
Stochastic process where consecutive increments are independent from the past.
Markov processes have the following property: given that its current state is known, the probability of any future event of the process is not altered by additional knowledge concerning its past behavior.
Maximum Entropy Mean-Field Theory Metastability Multifractal