Markov chain: Difference between revisions
CSV import Tags: mobile edit mobile web edit |
CSV import |
||
| Line 31: | Line 31: | ||
{{stub}} | {{stub}} | ||
{{dictionary-stub1}} | {{dictionary-stub1}} | ||
== Markov_chain == | |||
<gallery> | |||
File:Markovkate_01.svg|Markovkate 01 | |||
File:AAMarkov.jpg|A. A. Markov | |||
File:Intensities_vs_transition_probabilities.svg|Intensities vs transition probabilities | |||
File:PageRank_with_Markov_Chain.png|PageRank with Markov Chain | |||
</gallery> | |||
Latest revision as of 21:18, 23 February 2025
Markov chain is a stochastic process with the Markov property. The term "Markov chain" refers to the sequence of random variables such a process moves through, with the Markov property defining serial dependence only between the adjacent periods (as in a "chain"). It is named after the Russian mathematician Andrey Markov.
Definition[edit]
A Markov chain is a sequence of random variables X1, X2, X3, ... with the Markov property, namely that the probability of moving to the next state depends only on the present state and not on the previous states.
Properties[edit]
Markov chains have many properties, which are studied in topics such as stochastic processes, random walks, ergodic theory, and statistical mechanics. They are used as mathematical models of systems and processes in many fields.
Applications[edit]
Markov chains are used in various fields such as physics, chemistry, economics, social sciences, and engineering. They are particularly useful in the study of systems that follow a chain of linked events, which can be represented as states in a Markov chain.
See also[edit]
References[edit]
<references />



