In probability theory and statistics, a Markov chain or Markov process is a stochastic process describing a sequence of possible events in which the probability...
96 KB (12,900 words) - 21:01, 27 April 2025
In statistics, Markov chain Monte Carlo (MCMC) is a class of algorithms used to draw samples from a probability distribution. Given a probability distribution...
62 KB (8,537 words) - 04:54, 19 May 2025
A hidden Markov model (HMM) is a Markov model in which the observations are dependent on a latent (or hidden) Markov process (referred to as X {\displaystyle...
52 KB (6,811 words) - 04:08, 22 December 2024
examples of Markov chains and Markov processes in action. All examples are in the countable state space. For an overview of Markov chains in general state...
14 KB (2,405 words) - 17:52, 29 March 2025
A continuous-time Markov chain (CTMC) is a continuous stochastic process in which, for each state, the process will change state according to an exponential...
23 KB (4,240 words) - 18:35, 6 May 2025
In probability, a discrete-time Markov chain (DTMC) is a sequence of random variables, known as a stochastic process, in which the value of the next variable...
25 KB (4,252 words) - 18:52, 20 February 2025
In the mathematical theory of probability, an absorbing Markov chain is a Markov chain in which every state can reach an absorbing state. An absorbing...
12 KB (1,762 words) - 11:26, 30 December 2024
Markov chain geostatistics uses Markov chain spatial models, simulation algorithms and associated spatial correlation measures (e.g., transiogram) based...
2 KB (234 words) - 15:05, 12 September 2021
known as the Markov chain. He was also a strong, close to master-level, chess player. Markov and his younger brother Vladimir Andreyevich Markov (1871–1897)...
10 KB (1,072 words) - 15:39, 28 November 2024
In mathematics, the quantum Markov chain is a reformulation of the ideas of a classical Markov chain, replacing the classical definitions of probability...
2 KB (201 words) - 17:45, 26 February 2025
stochastic process satisfying the Markov property is known as a Markov chain. A stochastic process has the Markov property if the conditional probability...
8 KB (1,124 words) - 20:27, 8 March 2025
Subshift of finite type (redirect from Topological Markov chain)
probability measure on the set of subshifts. For example, consider the Markov chain given on the left on the states A , B 1 , B 2 {\displaystyle A,B_{1}...
16 KB (2,396 words) - 16:08, 20 December 2024
Markov chain on a measurable state space is a discrete-time-homogeneous Markov chain with a measurable space as state space. The definition of Markov...
5 KB (1,000 words) - 08:14, 16 October 2023
LZMA (redirect from Lempel-Ziv-Markov chain-Algorithm)
The Lempel–Ziv–Markov chain algorithm (LZMA) is an algorithm used to perform lossless data compression. It has been used in the 7z format of the 7-Zip...
31 KB (3,534 words) - 21:42, 4 May 2025
Gauss–Markov theorem Gauss–Markov process Markov blanket Markov boundary Markov chain Markov chain central limit theorem Additive Markov chain Markov additive...
2 KB (229 words) - 07:10, 17 June 2024
Stochastic matrix (redirect from Markov transition matrix)
stochastic matrix is a square matrix used to describe the transitions of a Markov chain. Each of its entries is a nonnegative real number representing a probability...
20 KB (2,959 words) - 14:55, 5 May 2025
In numerical methods for stochastic differential equations, the Markov chain approximation method (MCAM) belongs to the several numerical (schemes) approaches...
2 KB (225 words) - 13:20, 20 June 2017
from its connection to Markov chains, a concept developed by the Russian mathematician Andrey Markov. The "Markov" in "Markov decision process" refers...
35 KB (5,156 words) - 19:43, 21 March 2025
simplest Markov model is the Markov chain. It models the state of a system with a random variable that changes through time. In this context, the Markov property...
10 KB (1,231 words) - 22:12, 5 May 2025
In the mathematical theory of random processes, the Markov chain central limit theorem has a conclusion somewhat similar in form to that of the classic...
6 KB (1,166 words) - 16:29, 18 April 2025
Markov chain is the time until the Markov chain is "close" to its steady state distribution. More precisely, a fundamental result about Markov chains...
5 KB (604 words) - 20:16, 9 July 2024
mathematical theory of Markov chains, the Markov chain tree theorem is an expression for the stationary distribution of a Markov chain with finitely many...
4 KB (582 words) - 20:59, 14 April 2025
Detailed balance (redirect from Reversible markov chain)
balance in kinetics seem to be clear. A Markov process is called a reversible Markov process or reversible Markov chain if there exists a positive stationary...
35 KB (5,752 words) - 22:33, 12 April 2025
counting measures. The Markov chain is ergodic, so the shift example from above is a special case of the criterion. Markov chains with recurring communicating...
55 KB (8,917 words) - 23:47, 18 March 2025
probability theory, a telescoping Markov chain (TMC) is a vector-valued stochastic process that satisfies a Markov property and admits a hierarchical...
2 KB (401 words) - 18:52, 22 September 2024
characterize continuous-time Markov processes. In particular, they describe how the probability of a continuous-time Markov process in a certain state changes...
9 KB (1,438 words) - 22:49, 6 May 2025
M/G/1 queue (redirect from M/G/1 type Markov chain)
of jobs to the queue. Markov chains with generator matrices or block matrices of this form are called M/G/1 type Markov chains, a term coined by Marcel...
14 KB (1,787 words) - 08:05, 21 November 2024
boundary were coined by Judea Pearl in 1988. A Markov blanket can be constituted by a set of Markov chains. A Markov blanket of a random variable Y {\displaystyle...
4 KB (538 words) - 06:28, 15 May 2024
{\displaystyle S} in the Markov chain to letters in the alphabet Γ {\displaystyle \Gamma } . A unifilar Markov source is a Markov source for which the values...
2 KB (238 words) - 03:32, 13 March 2024
Random walk (section As a Markov chain)
) {\displaystyle O(a+b)} in the general one-dimensional random walk Markov chain. Some of the results mentioned above can be derived from properties of...
56 KB (7,703 words) - 01:28, 25 February 2025