When.com Web Search

Search results

  1. Results From The WOW.Com Content Network
  2. Continuous-time Markov chain - Wikipedia

    en.wikipedia.org/wiki/Continuous-time_Markov_chain

    A continuous-time Markov chain (CTMC) is a continuous stochastic process in which, for each state, the process will change state according to an exponential random variable and then move to a different state as specified by the probabilities of a stochastic matrix.

  3. Markov chain - Wikipedia

    en.wikipedia.org/wiki/Markov_chain

    Markov chains and continuous-time Markov processes are useful in chemistry when physical systems closely approximate the Markov property. For example, imagine a large number n of molecules in solution in state A, each of which can undergo a chemical reaction to state B with a certain average rate. Perhaps the molecule is an enzyme, and the ...

  4. Stopping time - Wikipedia

    en.wikipedia.org/wiki/Stopping_time

    Example of a stopping time: a hitting time of Brownian motion.The process starts at 0 and is stopped as soon as it hits 1. In probability theory, in particular in the study of stochastic processes, a stopping time (also Markov time, Markov moment, optional stopping time or optional time [1]) is a specific type of “random time”: a random variable whose value is interpreted as the time at ...

  5. Kolmogorov equations - Wikipedia

    en.wikipedia.org/wiki/Kolmogorov_equations

    Continuous-time Markov chains [ edit ] The original derivation of the equations by Kolmogorov starts with the Chapman–Kolmogorov equation (Kolmogorov called it fundamental equation ) for time-continuous and differentiable Markov processes on a finite, discrete state space. [ 2 ]

  6. Kemeny's constant - Wikipedia

    en.wikipedia.org/wiki/Kemeny's_constant

    In probability theory, Kemeny’s constant is the expected number of time steps required for a Markov chain to transition from a starting state i to a random destination state sampled from the Markov chain's stationary distribution.

  7. Examples of Markov chains - Wikipedia

    en.wikipedia.org/wiki/Examples_of_Markov_chains

    A game of snakes and ladders or any other game whose moves are determined entirely by dice is a Markov chain, indeed, an absorbing Markov chain. This is in contrast to card games such as blackjack, where the cards represent a 'memory' of the past moves. To see the difference, consider the probability for a certain event in the game.

  8. Markov chain mixing time - Wikipedia

    en.wikipedia.org/wiki/Markov_chain_mixing_time

    In probability theory, the mixing time of a Markov chain is the time until the Markov chain is "close" to its steady state distribution.. More precisely, a fundamental result about Markov chains is that a finite state irreducible aperiodic chain has a unique stationary distribution π and, regardless of the initial state, the time-t distribution of the chain converges to π as t tends to infinity.

  9. Infinitesimal generator (stochastic processes) - Wikipedia

    en.wikipedia.org/wiki/Infinitesimal_generator...

    In mathematics — specifically, in stochastic analysis — the infinitesimal generator of a Feller process (i.e. a continuous-time Markov process satisfying certain regularity conditions) is a Fourier multiplier operator [1] that encodes a great deal of information about the process.