Markovian

  1. Adjective.  (statistics, of a process) Exhibiting the Markov property, in which the conditional probability distribution of future states of the process, given the present state and all past states, depends only upon the present state and not on any past states.
  2. Adjective.  (alternative capitalization of, Markovian) .

This is an unmodified, but possibly outdated, definition from Wiktionary and used here under the Creative Commons license. Wiktionary is a great resource. If you like it too, please donate to Wikimedia.

This entry was last updated on RefTopia from its source on 3/20/2012.