Associations to the word «Markov»

Wiktionary

MARKOV, proper noun. A surname​.
MARKOV CHAIN, noun. (probability theory) A discrete-time stochastic process with the Markov property.
MARKOV CHAINS, noun. Plural of Markov chain
MARKOV JUMP PROCESS, noun. (mathematics) A time-dependent variable that starts in an initial state and stays in that state for a random time, when it makes a transition to another random state, and so on.
MARKOV PARTITION, noun. (math) A tool used in dynamical systems theory, allowing the methods of symbolic dynamics to be applied to the study of hyperbolic systems. By using a Markov partition, the system can be made to resemble a discrete-time Markov process, with the long-term dynamical characteristics of the system represented as a Markov shift.
MARKOV PARTITIONS, noun. Plural of Markov partition
MARKOV PROCESS, noun. (probability theory) A stochastic process in which the probability distribution of the current state is conditionally independent of the path of past states.
MARKOV PROCESSES, noun. Plural of Markov process

Dictionary definition

MARKOV, noun. Russian mathematician (1856-1922).

Wise words

In words, as fashions, the same rule will hold; Alike fantastic, if too new, or old: Be not the first by whom the new are tried, Nor yet the last to lay the old aside.
Alexander Pope