Markoff process

Markoff process
Markov process Mark"ov pro`cess, n. [after A. A. Markov, Russian mathematician, b. 1856, d. 1922.] (Statistics) a random process in which the probabilities of states in a series depend only on the properties of the immediately preceding state or the next preceeding state, independent of the path by which the preceding state was reached. It is distinguished from a {Markov chain} in that the states of a Markov process may be continuous as well as discrete. [Also spelled {Markoff process}.] [PJC]

The Collaborative International Dictionary of English. 2000.

Игры ⚽ Поможем решить контрольную работу

Look at other dictionaries:

  • Markoff process — noun a simple stochastic process in which the distribution of future states depends only on the present state and not on how it arrived in the present state • Syn: ↑Markov process • Hypernyms: ↑stochastic process • Hyponyms: ↑Markov chain,… …   Useful english dictionary

  • Markoff process — noun see Markov process …   New Collegiate Dictionary

  • Markoff chain — noun a Markov process for which the parameter is discrete time values • Syn: ↑Markov chain • Hypernyms: ↑Markov process, ↑Markoff process * * * noun see markov chain …   Useful english dictionary

  • Markoff chain — Markov chain Mark ov chain, n. [after A. A. Markov, Russian mathematician, b. 1856, d. 1922.] (Statistics) A random process (Markov process) in which the probabilities of discrete states in a series depend only on the properties of the… …   The Collaborative International Dictionary of English

  • Markov process — noun a simple stochastic process in which the distribution of future states depends only on the present state and not on how it arrived in the present state • Syn: ↑Markoff process • Hypernyms: ↑stochastic process • Hyponyms: ↑Markov chain,… …   Useful english dictionary

  • Markov process — Mark ov pro cess, n. [after A. A. Markov, Russian mathematician, b. 1856, d. 1922.] (Statistics) a random process in which the probabilities of states in a series depend only on the properties of the immediately preceding state or the next… …   The Collaborative International Dictionary of English

  • Markov process — Statistics. a process in which future values of a random variable are statistically determined by present events and dependent only on the event immediately preceding. Also, Markoff process. [1935 40; after Russian mathematician Andrei Andreevich …   Universalium

  • stochastic process — noun a statistical process involving a number of random variables depending on a variable parameter (which is usually time) • Hypernyms: ↑model, ↑theoretical account, ↑framework • Hyponyms: ↑Markov process, ↑Markoff process, ↑random walk, ↑ …   Useful english dictionary

  • Markov process — noun Date: 1938 a stochastic process (as Brownian motion) that resembles a Markov chain except that the states are continuous; also Markov chain called also Markoff process …   New Collegiate Dictionary

  • Markov process — [mär′kôf] n. a chain of random events in which only the present state influences the next future state, as in a genetic code: also Markoff process …   English World dictionary

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”